00:00:00.001 Started by upstream project "autotest-spdk-master-vs-dpdk-v22.11" build number 2229 00:00:00.001 originally caused by: 00:00:00.002 Started by upstream project "nightly-trigger" build number 3492 00:00:00.002 originally caused by: 00:00:00.002 Started by timer 00:00:00.002 Started by timer 00:00:00.019 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.020 The recommended git tool is: git 00:00:00.020 using credential 00000000-0000-0000-0000-000000000002 00:00:00.022 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.038 Fetching changes from the remote Git repository 00:00:00.040 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.057 Using shallow fetch with depth 1 00:00:00.057 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.057 > git --version # timeout=10 00:00:00.082 > git --version # 'git version 2.39.2' 00:00:00.083 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.123 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.123 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:02.293 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:02.306 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:02.317 Checking out Revision 7510e71a2b3ec6fca98e4ec196065590f900d444 (FETCH_HEAD) 00:00:02.317 > git config core.sparsecheckout # timeout=10 00:00:02.328 > git read-tree -mu HEAD # timeout=10 00:00:02.344 > git checkout -f 7510e71a2b3ec6fca98e4ec196065590f900d444 # timeout=5 00:00:02.360 Commit message: "kid: add issue 3541" 00:00:02.361 > git rev-list --no-walk 7510e71a2b3ec6fca98e4ec196065590f900d444 # timeout=10 00:00:02.445 [Pipeline] Start of Pipeline 00:00:02.463 [Pipeline] library 00:00:02.465 Loading library shm_lib@master 00:00:02.466 Library shm_lib@master is cached. Copying from home. 00:00:02.488 [Pipeline] node 00:00:02.503 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:02.505 [Pipeline] { 00:00:02.517 [Pipeline] catchError 00:00:02.519 [Pipeline] { 00:00:02.538 [Pipeline] wrap 00:00:02.550 [Pipeline] { 00:00:02.562 [Pipeline] stage 00:00:02.564 [Pipeline] { (Prologue) 00:00:02.589 [Pipeline] echo 00:00:02.590 Node: VM-host-WFP7 00:00:02.597 [Pipeline] cleanWs 00:00:02.608 [WS-CLEANUP] Deleting project workspace... 00:00:02.608 [WS-CLEANUP] Deferred wipeout is used... 00:00:02.615 [WS-CLEANUP] done 00:00:02.810 [Pipeline] setCustomBuildProperty 00:00:02.900 [Pipeline] httpRequest 00:00:03.300 [Pipeline] echo 00:00:03.301 Sorcerer 10.211.164.101 is alive 00:00:03.311 [Pipeline] retry 00:00:03.312 [Pipeline] { 00:00:03.328 [Pipeline] httpRequest 00:00:03.334 HttpMethod: GET 00:00:03.335 URL: http://10.211.164.101/packages/jbp_7510e71a2b3ec6fca98e4ec196065590f900d444.tar.gz 00:00:03.335 Sending request to url: http://10.211.164.101/packages/jbp_7510e71a2b3ec6fca98e4ec196065590f900d444.tar.gz 00:00:03.336 Response Code: HTTP/1.1 200 OK 00:00:03.336 Success: Status code 200 is in the accepted range: 200,404 00:00:03.337 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_7510e71a2b3ec6fca98e4ec196065590f900d444.tar.gz 00:00:03.481 [Pipeline] } 00:00:03.498 [Pipeline] // retry 00:00:03.504 [Pipeline] sh 00:00:03.789 + tar --no-same-owner -xf jbp_7510e71a2b3ec6fca98e4ec196065590f900d444.tar.gz 00:00:03.802 [Pipeline] httpRequest 00:00:04.185 [Pipeline] echo 00:00:04.186 Sorcerer 10.211.164.101 is alive 00:00:04.195 [Pipeline] retry 00:00:04.197 [Pipeline] { 00:00:04.209 [Pipeline] httpRequest 00:00:04.213 HttpMethod: GET 00:00:04.214 URL: http://10.211.164.101/packages/spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:04.214 Sending request to url: http://10.211.164.101/packages/spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:04.215 Response Code: HTTP/1.1 200 OK 00:00:04.215 Success: Status code 200 is in the accepted range: 200,404 00:00:04.216 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:21.392 [Pipeline] } 00:00:21.413 [Pipeline] // retry 00:00:21.422 [Pipeline] sh 00:00:21.709 + tar --no-same-owner -xf spdk_09cc66129742c68eb8ce46c42225a27c3c933a14.tar.gz 00:00:24.266 [Pipeline] sh 00:00:24.551 + git -C spdk log --oneline -n5 00:00:24.551 09cc66129 test/unit: add mixed busy/idle mock poller function in reactor_ut 00:00:24.551 a67b3561a dpdk: update submodule to include alarm_cancel fix 00:00:24.551 43f6d3385 nvmf: remove use of STAILQ for last_wqe events 00:00:24.551 9645421c5 nvmf: rename nvmf_rdma_qpair_process_ibv_event() 00:00:24.551 e6da32ee1 nvmf: rename nvmf_rdma_send_qpair_async_event() 00:00:24.572 [Pipeline] withCredentials 00:00:24.584 > git --version # timeout=10 00:00:24.598 > git --version # 'git version 2.39.2' 00:00:24.618 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:00:24.620 [Pipeline] { 00:00:24.630 [Pipeline] retry 00:00:24.632 [Pipeline] { 00:00:24.648 [Pipeline] sh 00:00:24.933 + git ls-remote http://dpdk.org/git/dpdk-stable v22.11.4 00:00:25.207 [Pipeline] } 00:00:25.228 [Pipeline] // retry 00:00:25.233 [Pipeline] } 00:00:25.252 [Pipeline] // withCredentials 00:00:25.263 [Pipeline] httpRequest 00:00:25.668 [Pipeline] echo 00:00:25.670 Sorcerer 10.211.164.101 is alive 00:00:25.678 [Pipeline] retry 00:00:25.680 [Pipeline] { 00:00:25.690 [Pipeline] httpRequest 00:00:25.693 HttpMethod: GET 00:00:25.694 URL: http://10.211.164.101/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:25.694 Sending request to url: http://10.211.164.101/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:00:25.706 Response Code: HTTP/1.1 200 OK 00:00:25.707 Success: Status code 200 is in the accepted range: 200,404 00:00:25.707 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:12.770 [Pipeline] } 00:01:12.839 [Pipeline] // retry 00:01:12.844 [Pipeline] sh 00:01:13.127 + tar --no-same-owner -xf dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:14.522 [Pipeline] sh 00:01:14.806 + git -C dpdk log --oneline -n5 00:01:14.806 caf0f5d395 version: 22.11.4 00:01:14.806 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:01:14.806 dc9c799c7d vhost: fix missing spinlock unlock 00:01:14.806 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:01:14.806 6ef77f2a5e net/gve: fix RX buffer size alignment 00:01:14.828 [Pipeline] writeFile 00:01:14.846 [Pipeline] sh 00:01:15.136 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:01:15.149 [Pipeline] sh 00:01:15.434 + cat autorun-spdk.conf 00:01:15.434 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:15.434 SPDK_RUN_ASAN=1 00:01:15.434 SPDK_RUN_UBSAN=1 00:01:15.434 SPDK_TEST_RAID=1 00:01:15.434 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:15.434 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:15.434 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:15.442 RUN_NIGHTLY=1 00:01:15.444 [Pipeline] } 00:01:15.459 [Pipeline] // stage 00:01:15.474 [Pipeline] stage 00:01:15.476 [Pipeline] { (Run VM) 00:01:15.489 [Pipeline] sh 00:01:15.774 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:01:15.774 + echo 'Start stage prepare_nvme.sh' 00:01:15.774 Start stage prepare_nvme.sh 00:01:15.774 + [[ -n 3 ]] 00:01:15.774 + disk_prefix=ex3 00:01:15.774 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:01:15.775 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:01:15.775 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:01:15.775 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:15.775 ++ SPDK_RUN_ASAN=1 00:01:15.775 ++ SPDK_RUN_UBSAN=1 00:01:15.775 ++ SPDK_TEST_RAID=1 00:01:15.775 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:15.775 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:15.775 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:15.775 ++ RUN_NIGHTLY=1 00:01:15.775 + cd /var/jenkins/workspace/raid-vg-autotest 00:01:15.775 + nvme_files=() 00:01:15.775 + declare -A nvme_files 00:01:15.775 + backend_dir=/var/lib/libvirt/images/backends 00:01:15.775 + nvme_files['nvme.img']=5G 00:01:15.775 + nvme_files['nvme-cmb.img']=5G 00:01:15.775 + nvme_files['nvme-multi0.img']=4G 00:01:15.775 + nvme_files['nvme-multi1.img']=4G 00:01:15.775 + nvme_files['nvme-multi2.img']=4G 00:01:15.775 + nvme_files['nvme-openstack.img']=8G 00:01:15.775 + nvme_files['nvme-zns.img']=5G 00:01:15.775 + (( SPDK_TEST_NVME_PMR == 1 )) 00:01:15.775 + (( SPDK_TEST_FTL == 1 )) 00:01:15.775 + (( SPDK_TEST_NVME_FDP == 1 )) 00:01:15.775 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:01:15.775 + for nvme in "${!nvme_files[@]}" 00:01:15.775 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-multi2.img -s 4G 00:01:15.775 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:01:15.775 + for nvme in "${!nvme_files[@]}" 00:01:15.775 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-cmb.img -s 5G 00:01:15.775 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:01:15.775 + for nvme in "${!nvme_files[@]}" 00:01:15.775 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-openstack.img -s 8G 00:01:15.775 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:01:15.775 + for nvme in "${!nvme_files[@]}" 00:01:15.775 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-zns.img -s 5G 00:01:15.775 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:01:15.775 + for nvme in "${!nvme_files[@]}" 00:01:15.775 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-multi1.img -s 4G 00:01:15.775 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:01:15.775 + for nvme in "${!nvme_files[@]}" 00:01:15.775 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme-multi0.img -s 4G 00:01:15.775 Formatting '/var/lib/libvirt/images/backends/ex3-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:01:15.775 + for nvme in "${!nvme_files[@]}" 00:01:15.775 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex3-nvme.img -s 5G 00:01:16.035 Formatting '/var/lib/libvirt/images/backends/ex3-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:01:16.035 ++ sudo grep -rl ex3-nvme.img /etc/libvirt/qemu 00:01:16.035 + echo 'End stage prepare_nvme.sh' 00:01:16.035 End stage prepare_nvme.sh 00:01:16.048 [Pipeline] sh 00:01:16.337 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:01:16.337 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex3-nvme.img -b /var/lib/libvirt/images/backends/ex3-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex3-nvme-multi1.img:/var/lib/libvirt/images/backends/ex3-nvme-multi2.img -H -a -v -f fedora39 00:01:16.337 00:01:16.337 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:01:16.337 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:01:16.337 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:01:16.337 HELP=0 00:01:16.337 DRY_RUN=0 00:01:16.337 NVME_FILE=/var/lib/libvirt/images/backends/ex3-nvme.img,/var/lib/libvirt/images/backends/ex3-nvme-multi0.img, 00:01:16.337 NVME_DISKS_TYPE=nvme,nvme, 00:01:16.337 NVME_AUTO_CREATE=0 00:01:16.337 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex3-nvme-multi1.img:/var/lib/libvirt/images/backends/ex3-nvme-multi2.img, 00:01:16.337 NVME_CMB=,, 00:01:16.337 NVME_PMR=,, 00:01:16.337 NVME_ZNS=,, 00:01:16.337 NVME_MS=,, 00:01:16.337 NVME_FDP=,, 00:01:16.337 SPDK_VAGRANT_DISTRO=fedora39 00:01:16.337 SPDK_VAGRANT_VMCPU=10 00:01:16.337 SPDK_VAGRANT_VMRAM=12288 00:01:16.337 SPDK_VAGRANT_PROVIDER=libvirt 00:01:16.337 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:01:16.337 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:01:16.337 SPDK_OPENSTACK_NETWORK=0 00:01:16.337 VAGRANT_PACKAGE_BOX=0 00:01:16.337 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:01:16.337 FORCE_DISTRO=true 00:01:16.337 VAGRANT_BOX_VERSION= 00:01:16.337 EXTRA_VAGRANTFILES= 00:01:16.337 NIC_MODEL=virtio 00:01:16.337 00:01:16.337 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:01:16.337 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:01:18.246 Bringing machine 'default' up with 'libvirt' provider... 00:01:18.507 ==> default: Creating image (snapshot of base box volume). 00:01:18.768 ==> default: Creating domain with the following settings... 00:01:18.768 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1727627980_13361f815673ce2326f6 00:01:18.768 ==> default: -- Domain type: kvm 00:01:18.768 ==> default: -- Cpus: 10 00:01:18.768 ==> default: -- Feature: acpi 00:01:18.768 ==> default: -- Feature: apic 00:01:18.768 ==> default: -- Feature: pae 00:01:18.768 ==> default: -- Memory: 12288M 00:01:18.768 ==> default: -- Memory Backing: hugepages: 00:01:18.768 ==> default: -- Management MAC: 00:01:18.768 ==> default: -- Loader: 00:01:18.768 ==> default: -- Nvram: 00:01:18.768 ==> default: -- Base box: spdk/fedora39 00:01:18.768 ==> default: -- Storage pool: default 00:01:18.768 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1727627980_13361f815673ce2326f6.img (20G) 00:01:18.768 ==> default: -- Volume Cache: default 00:01:18.768 ==> default: -- Kernel: 00:01:18.768 ==> default: -- Initrd: 00:01:18.768 ==> default: -- Graphics Type: vnc 00:01:18.768 ==> default: -- Graphics Port: -1 00:01:18.768 ==> default: -- Graphics IP: 127.0.0.1 00:01:18.768 ==> default: -- Graphics Password: Not defined 00:01:18.768 ==> default: -- Video Type: cirrus 00:01:18.768 ==> default: -- Video VRAM: 9216 00:01:18.768 ==> default: -- Sound Type: 00:01:18.768 ==> default: -- Keymap: en-us 00:01:18.768 ==> default: -- TPM Path: 00:01:18.768 ==> default: -- INPUT: type=mouse, bus=ps2 00:01:18.768 ==> default: -- Command line args: 00:01:18.768 ==> default: -> value=-device, 00:01:18.768 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:01:18.768 ==> default: -> value=-drive, 00:01:18.768 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme.img,if=none,id=nvme-0-drive0, 00:01:18.768 ==> default: -> value=-device, 00:01:18.768 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:18.768 ==> default: -> value=-device, 00:01:18.768 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:01:18.768 ==> default: -> value=-drive, 00:01:18.768 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:01:18.768 ==> default: -> value=-device, 00:01:18.768 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:18.768 ==> default: -> value=-drive, 00:01:18.768 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:01:18.768 ==> default: -> value=-device, 00:01:18.768 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:18.768 ==> default: -> value=-drive, 00:01:18.768 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex3-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:01:18.768 ==> default: -> value=-device, 00:01:18.768 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:18.768 ==> default: Creating shared folders metadata... 00:01:18.768 ==> default: Starting domain. 00:01:20.680 ==> default: Waiting for domain to get an IP address... 00:01:35.605 ==> default: Waiting for SSH to become available... 00:01:37.040 ==> default: Configuring and enabling network interfaces... 00:01:43.644 default: SSH address: 192.168.121.240:22 00:01:43.645 default: SSH username: vagrant 00:01:43.645 default: SSH auth method: private key 00:01:46.202 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:01:54.335 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:02:00.908 ==> default: Mounting SSHFS shared folder... 00:02:02.820 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:02:02.820 ==> default: Checking Mount.. 00:02:04.203 ==> default: Folder Successfully Mounted! 00:02:04.203 ==> default: Running provisioner: file... 00:02:05.142 default: ~/.gitconfig => .gitconfig 00:02:05.713 00:02:05.713 SUCCESS! 00:02:05.713 00:02:05.713 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:02:05.713 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:02:05.713 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:02:05.713 00:02:05.723 [Pipeline] } 00:02:05.738 [Pipeline] // stage 00:02:05.748 [Pipeline] dir 00:02:05.749 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:02:05.751 [Pipeline] { 00:02:05.765 [Pipeline] catchError 00:02:05.768 [Pipeline] { 00:02:05.783 [Pipeline] sh 00:02:06.068 + vagrant ssh-config --host vagrant 00:02:06.068 + sed -ne /^Host/,$p 00:02:06.068 + tee ssh_conf 00:02:08.605 Host vagrant 00:02:08.605 HostName 192.168.121.240 00:02:08.605 User vagrant 00:02:08.605 Port 22 00:02:08.605 UserKnownHostsFile /dev/null 00:02:08.605 StrictHostKeyChecking no 00:02:08.605 PasswordAuthentication no 00:02:08.605 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:02:08.605 IdentitiesOnly yes 00:02:08.605 LogLevel FATAL 00:02:08.605 ForwardAgent yes 00:02:08.605 ForwardX11 yes 00:02:08.605 00:02:08.620 [Pipeline] withEnv 00:02:08.622 [Pipeline] { 00:02:08.637 [Pipeline] sh 00:02:08.922 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:02:08.922 source /etc/os-release 00:02:08.922 [[ -e /image.version ]] && img=$(< /image.version) 00:02:08.922 # Minimal, systemd-like check. 00:02:08.922 if [[ -e /.dockerenv ]]; then 00:02:08.922 # Clear garbage from the node's name: 00:02:08.922 # agt-er_autotest_547-896 -> autotest_547-896 00:02:08.922 # $HOSTNAME is the actual container id 00:02:08.922 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:02:08.922 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:02:08.922 # We can assume this is a mount from a host where container is running, 00:02:08.922 # so fetch its hostname to easily identify the target swarm worker. 00:02:08.922 container="$(< /etc/hostname) ($agent)" 00:02:08.922 else 00:02:08.922 # Fallback 00:02:08.922 container=$agent 00:02:08.922 fi 00:02:08.922 fi 00:02:08.922 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:02:08.922 00:02:09.195 [Pipeline] } 00:02:09.214 [Pipeline] // withEnv 00:02:09.224 [Pipeline] setCustomBuildProperty 00:02:09.241 [Pipeline] stage 00:02:09.244 [Pipeline] { (Tests) 00:02:09.264 [Pipeline] sh 00:02:09.550 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:02:09.824 [Pipeline] sh 00:02:10.108 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:02:10.384 [Pipeline] timeout 00:02:10.385 Timeout set to expire in 1 hr 30 min 00:02:10.387 [Pipeline] { 00:02:10.401 [Pipeline] sh 00:02:10.685 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:02:11.255 HEAD is now at 09cc66129 test/unit: add mixed busy/idle mock poller function in reactor_ut 00:02:11.268 [Pipeline] sh 00:02:11.551 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:02:11.825 [Pipeline] sh 00:02:12.108 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:02:12.389 [Pipeline] sh 00:02:12.751 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:02:13.011 ++ readlink -f spdk_repo 00:02:13.011 + DIR_ROOT=/home/vagrant/spdk_repo 00:02:13.011 + [[ -n /home/vagrant/spdk_repo ]] 00:02:13.011 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:02:13.011 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:02:13.011 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:02:13.011 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:02:13.011 + [[ -d /home/vagrant/spdk_repo/output ]] 00:02:13.011 + [[ raid-vg-autotest == pkgdep-* ]] 00:02:13.011 + cd /home/vagrant/spdk_repo 00:02:13.011 + source /etc/os-release 00:02:13.011 ++ NAME='Fedora Linux' 00:02:13.011 ++ VERSION='39 (Cloud Edition)' 00:02:13.011 ++ ID=fedora 00:02:13.011 ++ VERSION_ID=39 00:02:13.011 ++ VERSION_CODENAME= 00:02:13.011 ++ PLATFORM_ID=platform:f39 00:02:13.011 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:02:13.011 ++ ANSI_COLOR='0;38;2;60;110;180' 00:02:13.011 ++ LOGO=fedora-logo-icon 00:02:13.011 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:02:13.011 ++ HOME_URL=https://fedoraproject.org/ 00:02:13.011 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:02:13.011 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:02:13.011 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:02:13.011 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:02:13.011 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:02:13.011 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:02:13.011 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:02:13.011 ++ SUPPORT_END=2024-11-12 00:02:13.011 ++ VARIANT='Cloud Edition' 00:02:13.011 ++ VARIANT_ID=cloud 00:02:13.011 + uname -a 00:02:13.011 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:02:13.011 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:02:13.581 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:02:13.581 Hugepages 00:02:13.581 node hugesize free / total 00:02:13.581 node0 1048576kB 0 / 0 00:02:13.581 node0 2048kB 0 / 0 00:02:13.581 00:02:13.581 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:13.581 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:02:13.581 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme1 nvme1n1 00:02:13.581 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme0 nvme0n1 nvme0n2 nvme0n3 00:02:13.581 + rm -f /tmp/spdk-ld-path 00:02:13.581 + source autorun-spdk.conf 00:02:13.581 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:13.581 ++ SPDK_RUN_ASAN=1 00:02:13.581 ++ SPDK_RUN_UBSAN=1 00:02:13.581 ++ SPDK_TEST_RAID=1 00:02:13.581 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:13.581 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:13.581 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:13.581 ++ RUN_NIGHTLY=1 00:02:13.581 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:02:13.581 + [[ -n '' ]] 00:02:13.581 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:02:13.841 + for M in /var/spdk/build-*-manifest.txt 00:02:13.841 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:02:13.841 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:13.841 + for M in /var/spdk/build-*-manifest.txt 00:02:13.841 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:02:13.841 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:13.841 + for M in /var/spdk/build-*-manifest.txt 00:02:13.841 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:02:13.841 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:13.841 ++ uname 00:02:13.841 + [[ Linux == \L\i\n\u\x ]] 00:02:13.841 + sudo dmesg -T 00:02:13.841 + sudo dmesg --clear 00:02:13.841 + dmesg_pid=6155 00:02:13.841 + [[ Fedora Linux == FreeBSD ]] 00:02:13.841 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:13.841 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:13.841 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:02:13.841 + [[ -x /usr/src/fio-static/fio ]] 00:02:13.841 + sudo dmesg -Tw 00:02:13.841 + export FIO_BIN=/usr/src/fio-static/fio 00:02:13.841 + FIO_BIN=/usr/src/fio-static/fio 00:02:13.841 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:02:13.841 + [[ ! -v VFIO_QEMU_BIN ]] 00:02:13.841 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:02:13.841 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:13.841 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:13.841 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:02:13.841 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:13.841 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:13.841 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:13.841 Test configuration: 00:02:13.841 SPDK_RUN_FUNCTIONAL_TEST=1 00:02:13.841 SPDK_RUN_ASAN=1 00:02:13.841 SPDK_RUN_UBSAN=1 00:02:13.841 SPDK_TEST_RAID=1 00:02:13.841 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:13.841 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:13.841 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:14.102 RUN_NIGHTLY=1 16:40:35 -- common/autotest_common.sh@1680 -- $ [[ n == y ]] 00:02:14.102 16:40:35 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:02:14.102 16:40:35 -- scripts/common.sh@15 -- $ shopt -s extglob 00:02:14.102 16:40:35 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:02:14.102 16:40:35 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:14.102 16:40:35 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:14.102 16:40:35 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:14.102 16:40:35 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:14.102 16:40:35 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:14.102 16:40:35 -- paths/export.sh@5 -- $ export PATH 00:02:14.102 16:40:35 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:14.102 16:40:35 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:02:14.102 16:40:35 -- common/autobuild_common.sh@479 -- $ date +%s 00:02:14.102 16:40:35 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1727628035.XXXXXX 00:02:14.102 16:40:35 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1727628035.nS5YpC 00:02:14.102 16:40:35 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:02:14.102 16:40:35 -- common/autobuild_common.sh@485 -- $ '[' -n v22.11.4 ']' 00:02:14.102 16:40:35 -- common/autobuild_common.sh@486 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:14.102 16:40:35 -- common/autobuild_common.sh@486 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:02:14.102 16:40:35 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:02:14.102 16:40:35 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:02:14.102 16:40:35 -- common/autobuild_common.sh@495 -- $ get_config_params 00:02:14.102 16:40:35 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:02:14.102 16:40:35 -- common/autotest_common.sh@10 -- $ set +x 00:02:14.102 16:40:35 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:02:14.102 16:40:35 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:02:14.102 16:40:35 -- pm/common@17 -- $ local monitor 00:02:14.102 16:40:35 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:14.102 16:40:35 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:14.102 16:40:35 -- pm/common@25 -- $ sleep 1 00:02:14.102 16:40:35 -- pm/common@21 -- $ date +%s 00:02:14.102 16:40:35 -- pm/common@21 -- $ date +%s 00:02:14.102 16:40:35 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1727628035 00:02:14.102 16:40:35 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1727628035 00:02:14.102 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1727628035_collect-cpu-load.pm.log 00:02:14.102 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1727628035_collect-vmstat.pm.log 00:02:15.042 16:40:36 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:02:15.042 16:40:36 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:02:15.042 16:40:36 -- spdk/autobuild.sh@12 -- $ umask 022 00:02:15.042 16:40:36 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:15.042 16:40:36 -- spdk/autobuild.sh@16 -- $ date -u 00:02:15.042 Sun Sep 29 04:40:36 PM UTC 2024 00:02:15.042 16:40:36 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:02:15.042 v25.01-pre-17-g09cc66129 00:02:15.042 16:40:36 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:02:15.042 16:40:36 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:02:15.042 16:40:36 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:02:15.042 16:40:36 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:15.042 16:40:36 -- common/autotest_common.sh@10 -- $ set +x 00:02:15.042 ************************************ 00:02:15.042 START TEST asan 00:02:15.042 ************************************ 00:02:15.042 using asan 00:02:15.042 16:40:36 asan -- common/autotest_common.sh@1125 -- $ echo 'using asan' 00:02:15.042 00:02:15.042 real 0m0.000s 00:02:15.042 user 0m0.000s 00:02:15.042 sys 0m0.000s 00:02:15.042 16:40:36 asan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:15.042 16:40:36 asan -- common/autotest_common.sh@10 -- $ set +x 00:02:15.042 ************************************ 00:02:15.042 END TEST asan 00:02:15.042 ************************************ 00:02:15.042 16:40:36 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:02:15.042 16:40:36 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:02:15.042 16:40:36 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:02:15.042 16:40:36 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:15.042 16:40:36 -- common/autotest_common.sh@10 -- $ set +x 00:02:15.042 ************************************ 00:02:15.042 START TEST ubsan 00:02:15.042 ************************************ 00:02:15.042 using ubsan 00:02:15.042 16:40:36 ubsan -- common/autotest_common.sh@1125 -- $ echo 'using ubsan' 00:02:15.042 00:02:15.042 real 0m0.001s 00:02:15.042 user 0m0.000s 00:02:15.042 sys 0m0.000s 00:02:15.042 16:40:36 ubsan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:15.042 16:40:36 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:02:15.042 ************************************ 00:02:15.042 END TEST ubsan 00:02:15.042 ************************************ 00:02:15.302 16:40:36 -- spdk/autobuild.sh@27 -- $ '[' -n v22.11.4 ']' 00:02:15.302 16:40:36 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:02:15.302 16:40:36 -- common/autobuild_common.sh@442 -- $ run_test build_native_dpdk _build_native_dpdk 00:02:15.302 16:40:36 -- common/autotest_common.sh@1101 -- $ '[' 2 -le 1 ']' 00:02:15.302 16:40:36 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:15.302 16:40:36 -- common/autotest_common.sh@10 -- $ set +x 00:02:15.302 ************************************ 00:02:15.302 START TEST build_native_dpdk 00:02:15.302 ************************************ 00:02:15.302 16:40:36 build_native_dpdk -- common/autotest_common.sh@1125 -- $ _build_native_dpdk 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:02:15.302 caf0f5d395 version: 22.11.4 00:02:15.302 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:02:15.302 dc9c799c7d vhost: fix missing spinlock unlock 00:02:15.302 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:02:15.302 6ef77f2a5e net/gve: fix RX buffer size alignment 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=22.11.4 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@100 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base") 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@102 -- $ local mlx5_libs_added=n 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@139 -- $ [[ 0 -eq 1 ]] 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@167 -- $ cd /home/vagrant/spdk_repo/dpdk 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@168 -- $ uname -s 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@168 -- $ '[' Linux = Linux ']' 00:02:15.302 16:40:36 build_native_dpdk -- common/autobuild_common.sh@169 -- $ lt 22.11.4 21.11.0 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 21.11.0 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:02:15.302 16:40:36 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@173 -- $ patch -p1 00:02:15.303 patching file config/rte_config.h 00:02:15.303 Hunk #1 succeeded at 60 (offset 1 line). 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 22.11.4 24.07.0 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 24.07.0 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@177 -- $ patch -p1 00:02:15.303 patching file lib/pcapng/rte_pcapng.c 00:02:15.303 Hunk #1 succeeded at 110 (offset -18 lines). 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@179 -- $ ge 22.11.4 24.07.0 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 22.11.4 '>=' 24.07.0 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:15.303 16:40:36 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@183 -- $ dpdk_kmods=false 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@184 -- $ uname -s 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@184 -- $ '[' Linux = FreeBSD ']' 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@188 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base 00:02:15.303 16:40:36 build_native_dpdk -- common/autobuild_common.sh@188 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:21.876 The Meson build system 00:02:21.876 Version: 1.5.0 00:02:21.876 Source dir: /home/vagrant/spdk_repo/dpdk 00:02:21.876 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:02:21.876 Build type: native build 00:02:21.876 Program cat found: YES (/usr/bin/cat) 00:02:21.876 Project name: DPDK 00:02:21.876 Project version: 22.11.4 00:02:21.876 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:21.876 C linker for the host machine: gcc ld.bfd 2.40-14 00:02:21.876 Host machine cpu family: x86_64 00:02:21.876 Host machine cpu: x86_64 00:02:21.876 Message: ## Building in Developer Mode ## 00:02:21.876 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:21.876 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:02:21.876 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:02:21.876 Program objdump found: YES (/usr/bin/objdump) 00:02:21.876 Program python3 found: YES (/usr/bin/python3) 00:02:21.876 Program cat found: YES (/usr/bin/cat) 00:02:21.876 config/meson.build:83: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:02:21.876 Checking for size of "void *" : 8 00:02:21.876 Checking for size of "void *" : 8 (cached) 00:02:21.876 Library m found: YES 00:02:21.876 Library numa found: YES 00:02:21.876 Has header "numaif.h" : YES 00:02:21.876 Library fdt found: NO 00:02:21.876 Library execinfo found: NO 00:02:21.876 Has header "execinfo.h" : YES 00:02:21.876 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:21.876 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:21.876 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:21.876 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:21.876 Run-time dependency openssl found: YES 3.1.1 00:02:21.876 Run-time dependency libpcap found: YES 1.10.4 00:02:21.876 Has header "pcap.h" with dependency libpcap: YES 00:02:21.876 Compiler for C supports arguments -Wcast-qual: YES 00:02:21.876 Compiler for C supports arguments -Wdeprecated: YES 00:02:21.876 Compiler for C supports arguments -Wformat: YES 00:02:21.876 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:21.876 Compiler for C supports arguments -Wformat-security: NO 00:02:21.876 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:21.876 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:21.876 Compiler for C supports arguments -Wnested-externs: YES 00:02:21.876 Compiler for C supports arguments -Wold-style-definition: YES 00:02:21.876 Compiler for C supports arguments -Wpointer-arith: YES 00:02:21.876 Compiler for C supports arguments -Wsign-compare: YES 00:02:21.876 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:21.876 Compiler for C supports arguments -Wundef: YES 00:02:21.876 Compiler for C supports arguments -Wwrite-strings: YES 00:02:21.876 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:21.876 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:21.876 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:21.876 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:21.876 Compiler for C supports arguments -mavx512f: YES 00:02:21.876 Checking if "AVX512 checking" compiles: YES 00:02:21.876 Fetching value of define "__SSE4_2__" : 1 00:02:21.876 Fetching value of define "__AES__" : 1 00:02:21.876 Fetching value of define "__AVX__" : 1 00:02:21.876 Fetching value of define "__AVX2__" : 1 00:02:21.876 Fetching value of define "__AVX512BW__" : 1 00:02:21.876 Fetching value of define "__AVX512CD__" : 1 00:02:21.876 Fetching value of define "__AVX512DQ__" : 1 00:02:21.876 Fetching value of define "__AVX512F__" : 1 00:02:21.876 Fetching value of define "__AVX512VL__" : 1 00:02:21.876 Fetching value of define "__PCLMUL__" : 1 00:02:21.876 Fetching value of define "__RDRND__" : 1 00:02:21.876 Fetching value of define "__RDSEED__" : 1 00:02:21.876 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:21.876 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:21.876 Message: lib/kvargs: Defining dependency "kvargs" 00:02:21.876 Message: lib/telemetry: Defining dependency "telemetry" 00:02:21.876 Checking for function "getentropy" : YES 00:02:21.876 Message: lib/eal: Defining dependency "eal" 00:02:21.876 Message: lib/ring: Defining dependency "ring" 00:02:21.876 Message: lib/rcu: Defining dependency "rcu" 00:02:21.876 Message: lib/mempool: Defining dependency "mempool" 00:02:21.876 Message: lib/mbuf: Defining dependency "mbuf" 00:02:21.876 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:21.876 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:21.876 Compiler for C supports arguments -mpclmul: YES 00:02:21.876 Compiler for C supports arguments -maes: YES 00:02:21.876 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:21.876 Compiler for C supports arguments -mavx512bw: YES 00:02:21.876 Compiler for C supports arguments -mavx512dq: YES 00:02:21.876 Compiler for C supports arguments -mavx512vl: YES 00:02:21.876 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:21.876 Compiler for C supports arguments -mavx2: YES 00:02:21.876 Compiler for C supports arguments -mavx: YES 00:02:21.876 Message: lib/net: Defining dependency "net" 00:02:21.876 Message: lib/meter: Defining dependency "meter" 00:02:21.876 Message: lib/ethdev: Defining dependency "ethdev" 00:02:21.876 Message: lib/pci: Defining dependency "pci" 00:02:21.876 Message: lib/cmdline: Defining dependency "cmdline" 00:02:21.876 Message: lib/metrics: Defining dependency "metrics" 00:02:21.876 Message: lib/hash: Defining dependency "hash" 00:02:21.876 Message: lib/timer: Defining dependency "timer" 00:02:21.876 Fetching value of define "__AVX2__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512CD__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:21.876 Message: lib/acl: Defining dependency "acl" 00:02:21.876 Message: lib/bbdev: Defining dependency "bbdev" 00:02:21.876 Message: lib/bitratestats: Defining dependency "bitratestats" 00:02:21.876 Run-time dependency libelf found: YES 0.191 00:02:21.876 Message: lib/bpf: Defining dependency "bpf" 00:02:21.876 Message: lib/cfgfile: Defining dependency "cfgfile" 00:02:21.876 Message: lib/compressdev: Defining dependency "compressdev" 00:02:21.876 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:21.876 Message: lib/distributor: Defining dependency "distributor" 00:02:21.876 Message: lib/efd: Defining dependency "efd" 00:02:21.876 Message: lib/eventdev: Defining dependency "eventdev" 00:02:21.876 Message: lib/gpudev: Defining dependency "gpudev" 00:02:21.876 Message: lib/gro: Defining dependency "gro" 00:02:21.876 Message: lib/gso: Defining dependency "gso" 00:02:21.876 Message: lib/ip_frag: Defining dependency "ip_frag" 00:02:21.876 Message: lib/jobstats: Defining dependency "jobstats" 00:02:21.876 Message: lib/latencystats: Defining dependency "latencystats" 00:02:21.876 Message: lib/lpm: Defining dependency "lpm" 00:02:21.876 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512IFMA__" : (undefined) 00:02:21.876 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:02:21.876 Message: lib/member: Defining dependency "member" 00:02:21.876 Message: lib/pcapng: Defining dependency "pcapng" 00:02:21.876 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:21.876 Message: lib/power: Defining dependency "power" 00:02:21.876 Message: lib/rawdev: Defining dependency "rawdev" 00:02:21.876 Message: lib/regexdev: Defining dependency "regexdev" 00:02:21.876 Message: lib/dmadev: Defining dependency "dmadev" 00:02:21.876 Message: lib/rib: Defining dependency "rib" 00:02:21.876 Message: lib/reorder: Defining dependency "reorder" 00:02:21.876 Message: lib/sched: Defining dependency "sched" 00:02:21.876 Message: lib/security: Defining dependency "security" 00:02:21.876 Message: lib/stack: Defining dependency "stack" 00:02:21.876 Has header "linux/userfaultfd.h" : YES 00:02:21.876 Message: lib/vhost: Defining dependency "vhost" 00:02:21.876 Message: lib/ipsec: Defining dependency "ipsec" 00:02:21.876 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:21.876 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:21.876 Message: lib/fib: Defining dependency "fib" 00:02:21.876 Message: lib/port: Defining dependency "port" 00:02:21.876 Message: lib/pdump: Defining dependency "pdump" 00:02:21.876 Message: lib/table: Defining dependency "table" 00:02:21.876 Message: lib/pipeline: Defining dependency "pipeline" 00:02:21.876 Message: lib/graph: Defining dependency "graph" 00:02:21.876 Message: lib/node: Defining dependency "node" 00:02:21.877 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:21.877 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:21.877 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:21.877 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:21.877 Compiler for C supports arguments -Wno-sign-compare: YES 00:02:21.877 Compiler for C supports arguments -Wno-unused-value: YES 00:02:21.877 Compiler for C supports arguments -Wno-format: YES 00:02:21.877 Compiler for C supports arguments -Wno-format-security: YES 00:02:21.877 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:02:21.877 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:02:22.138 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:02:22.138 Compiler for C supports arguments -Wno-unused-parameter: YES 00:02:22.138 Fetching value of define "__AVX2__" : 1 (cached) 00:02:22.138 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:22.138 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:22.138 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:22.138 Compiler for C supports arguments -mavx512bw: YES (cached) 00:02:22.138 Compiler for C supports arguments -march=skylake-avx512: YES 00:02:22.138 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:02:22.138 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:22.138 Configuring doxy-api.conf using configuration 00:02:22.138 Program sphinx-build found: NO 00:02:22.138 Configuring rte_build_config.h using configuration 00:02:22.138 Message: 00:02:22.138 ================= 00:02:22.138 Applications Enabled 00:02:22.138 ================= 00:02:22.138 00:02:22.138 apps: 00:02:22.138 dumpcap, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, test-crypto-perf, 00:02:22.138 test-eventdev, test-fib, test-flow-perf, test-gpudev, test-pipeline, test-pmd, test-regex, test-sad, 00:02:22.138 test-security-perf, 00:02:22.138 00:02:22.138 Message: 00:02:22.138 ================= 00:02:22.138 Libraries Enabled 00:02:22.138 ================= 00:02:22.138 00:02:22.138 libs: 00:02:22.138 kvargs, telemetry, eal, ring, rcu, mempool, mbuf, net, 00:02:22.138 meter, ethdev, pci, cmdline, metrics, hash, timer, acl, 00:02:22.138 bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, efd, 00:02:22.138 eventdev, gpudev, gro, gso, ip_frag, jobstats, latencystats, lpm, 00:02:22.138 member, pcapng, power, rawdev, regexdev, dmadev, rib, reorder, 00:02:22.138 sched, security, stack, vhost, ipsec, fib, port, pdump, 00:02:22.138 table, pipeline, graph, node, 00:02:22.138 00:02:22.138 Message: 00:02:22.138 =============== 00:02:22.138 Drivers Enabled 00:02:22.138 =============== 00:02:22.138 00:02:22.138 common: 00:02:22.138 00:02:22.138 bus: 00:02:22.138 pci, vdev, 00:02:22.138 mempool: 00:02:22.138 ring, 00:02:22.138 dma: 00:02:22.138 00:02:22.138 net: 00:02:22.138 i40e, 00:02:22.138 raw: 00:02:22.138 00:02:22.138 crypto: 00:02:22.138 00:02:22.138 compress: 00:02:22.138 00:02:22.138 regex: 00:02:22.138 00:02:22.138 vdpa: 00:02:22.138 00:02:22.138 event: 00:02:22.138 00:02:22.138 baseband: 00:02:22.138 00:02:22.138 gpu: 00:02:22.138 00:02:22.138 00:02:22.138 Message: 00:02:22.138 ================= 00:02:22.138 Content Skipped 00:02:22.138 ================= 00:02:22.138 00:02:22.138 apps: 00:02:22.138 00:02:22.138 libs: 00:02:22.138 kni: explicitly disabled via build config (deprecated lib) 00:02:22.138 flow_classify: explicitly disabled via build config (deprecated lib) 00:02:22.138 00:02:22.138 drivers: 00:02:22.138 common/cpt: not in enabled drivers build config 00:02:22.138 common/dpaax: not in enabled drivers build config 00:02:22.138 common/iavf: not in enabled drivers build config 00:02:22.138 common/idpf: not in enabled drivers build config 00:02:22.138 common/mvep: not in enabled drivers build config 00:02:22.138 common/octeontx: not in enabled drivers build config 00:02:22.138 bus/auxiliary: not in enabled drivers build config 00:02:22.138 bus/dpaa: not in enabled drivers build config 00:02:22.138 bus/fslmc: not in enabled drivers build config 00:02:22.138 bus/ifpga: not in enabled drivers build config 00:02:22.138 bus/vmbus: not in enabled drivers build config 00:02:22.138 common/cnxk: not in enabled drivers build config 00:02:22.138 common/mlx5: not in enabled drivers build config 00:02:22.138 common/qat: not in enabled drivers build config 00:02:22.138 common/sfc_efx: not in enabled drivers build config 00:02:22.138 mempool/bucket: not in enabled drivers build config 00:02:22.138 mempool/cnxk: not in enabled drivers build config 00:02:22.138 mempool/dpaa: not in enabled drivers build config 00:02:22.138 mempool/dpaa2: not in enabled drivers build config 00:02:22.138 mempool/octeontx: not in enabled drivers build config 00:02:22.138 mempool/stack: not in enabled drivers build config 00:02:22.138 dma/cnxk: not in enabled drivers build config 00:02:22.138 dma/dpaa: not in enabled drivers build config 00:02:22.138 dma/dpaa2: not in enabled drivers build config 00:02:22.138 dma/hisilicon: not in enabled drivers build config 00:02:22.138 dma/idxd: not in enabled drivers build config 00:02:22.138 dma/ioat: not in enabled drivers build config 00:02:22.138 dma/skeleton: not in enabled drivers build config 00:02:22.138 net/af_packet: not in enabled drivers build config 00:02:22.138 net/af_xdp: not in enabled drivers build config 00:02:22.138 net/ark: not in enabled drivers build config 00:02:22.138 net/atlantic: not in enabled drivers build config 00:02:22.138 net/avp: not in enabled drivers build config 00:02:22.138 net/axgbe: not in enabled drivers build config 00:02:22.138 net/bnx2x: not in enabled drivers build config 00:02:22.138 net/bnxt: not in enabled drivers build config 00:02:22.138 net/bonding: not in enabled drivers build config 00:02:22.138 net/cnxk: not in enabled drivers build config 00:02:22.138 net/cxgbe: not in enabled drivers build config 00:02:22.138 net/dpaa: not in enabled drivers build config 00:02:22.138 net/dpaa2: not in enabled drivers build config 00:02:22.138 net/e1000: not in enabled drivers build config 00:02:22.138 net/ena: not in enabled drivers build config 00:02:22.138 net/enetc: not in enabled drivers build config 00:02:22.138 net/enetfec: not in enabled drivers build config 00:02:22.139 net/enic: not in enabled drivers build config 00:02:22.139 net/failsafe: not in enabled drivers build config 00:02:22.139 net/fm10k: not in enabled drivers build config 00:02:22.139 net/gve: not in enabled drivers build config 00:02:22.139 net/hinic: not in enabled drivers build config 00:02:22.139 net/hns3: not in enabled drivers build config 00:02:22.139 net/iavf: not in enabled drivers build config 00:02:22.139 net/ice: not in enabled drivers build config 00:02:22.139 net/idpf: not in enabled drivers build config 00:02:22.139 net/igc: not in enabled drivers build config 00:02:22.139 net/ionic: not in enabled drivers build config 00:02:22.139 net/ipn3ke: not in enabled drivers build config 00:02:22.139 net/ixgbe: not in enabled drivers build config 00:02:22.139 net/kni: not in enabled drivers build config 00:02:22.139 net/liquidio: not in enabled drivers build config 00:02:22.139 net/mana: not in enabled drivers build config 00:02:22.139 net/memif: not in enabled drivers build config 00:02:22.139 net/mlx4: not in enabled drivers build config 00:02:22.139 net/mlx5: not in enabled drivers build config 00:02:22.139 net/mvneta: not in enabled drivers build config 00:02:22.139 net/mvpp2: not in enabled drivers build config 00:02:22.139 net/netvsc: not in enabled drivers build config 00:02:22.139 net/nfb: not in enabled drivers build config 00:02:22.139 net/nfp: not in enabled drivers build config 00:02:22.139 net/ngbe: not in enabled drivers build config 00:02:22.139 net/null: not in enabled drivers build config 00:02:22.139 net/octeontx: not in enabled drivers build config 00:02:22.139 net/octeon_ep: not in enabled drivers build config 00:02:22.139 net/pcap: not in enabled drivers build config 00:02:22.139 net/pfe: not in enabled drivers build config 00:02:22.139 net/qede: not in enabled drivers build config 00:02:22.139 net/ring: not in enabled drivers build config 00:02:22.139 net/sfc: not in enabled drivers build config 00:02:22.139 net/softnic: not in enabled drivers build config 00:02:22.139 net/tap: not in enabled drivers build config 00:02:22.139 net/thunderx: not in enabled drivers build config 00:02:22.139 net/txgbe: not in enabled drivers build config 00:02:22.139 net/vdev_netvsc: not in enabled drivers build config 00:02:22.139 net/vhost: not in enabled drivers build config 00:02:22.139 net/virtio: not in enabled drivers build config 00:02:22.139 net/vmxnet3: not in enabled drivers build config 00:02:22.139 raw/cnxk_bphy: not in enabled drivers build config 00:02:22.139 raw/cnxk_gpio: not in enabled drivers build config 00:02:22.139 raw/dpaa2_cmdif: not in enabled drivers build config 00:02:22.139 raw/ifpga: not in enabled drivers build config 00:02:22.139 raw/ntb: not in enabled drivers build config 00:02:22.139 raw/skeleton: not in enabled drivers build config 00:02:22.139 crypto/armv8: not in enabled drivers build config 00:02:22.139 crypto/bcmfs: not in enabled drivers build config 00:02:22.139 crypto/caam_jr: not in enabled drivers build config 00:02:22.139 crypto/ccp: not in enabled drivers build config 00:02:22.139 crypto/cnxk: not in enabled drivers build config 00:02:22.139 crypto/dpaa_sec: not in enabled drivers build config 00:02:22.139 crypto/dpaa2_sec: not in enabled drivers build config 00:02:22.139 crypto/ipsec_mb: not in enabled drivers build config 00:02:22.139 crypto/mlx5: not in enabled drivers build config 00:02:22.139 crypto/mvsam: not in enabled drivers build config 00:02:22.139 crypto/nitrox: not in enabled drivers build config 00:02:22.139 crypto/null: not in enabled drivers build config 00:02:22.139 crypto/octeontx: not in enabled drivers build config 00:02:22.139 crypto/openssl: not in enabled drivers build config 00:02:22.139 crypto/scheduler: not in enabled drivers build config 00:02:22.139 crypto/uadk: not in enabled drivers build config 00:02:22.139 crypto/virtio: not in enabled drivers build config 00:02:22.139 compress/isal: not in enabled drivers build config 00:02:22.139 compress/mlx5: not in enabled drivers build config 00:02:22.139 compress/octeontx: not in enabled drivers build config 00:02:22.139 compress/zlib: not in enabled drivers build config 00:02:22.139 regex/mlx5: not in enabled drivers build config 00:02:22.139 regex/cn9k: not in enabled drivers build config 00:02:22.139 vdpa/ifc: not in enabled drivers build config 00:02:22.139 vdpa/mlx5: not in enabled drivers build config 00:02:22.139 vdpa/sfc: not in enabled drivers build config 00:02:22.139 event/cnxk: not in enabled drivers build config 00:02:22.139 event/dlb2: not in enabled drivers build config 00:02:22.139 event/dpaa: not in enabled drivers build config 00:02:22.139 event/dpaa2: not in enabled drivers build config 00:02:22.139 event/dsw: not in enabled drivers build config 00:02:22.139 event/opdl: not in enabled drivers build config 00:02:22.139 event/skeleton: not in enabled drivers build config 00:02:22.139 event/sw: not in enabled drivers build config 00:02:22.139 event/octeontx: not in enabled drivers build config 00:02:22.139 baseband/acc: not in enabled drivers build config 00:02:22.139 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:02:22.139 baseband/fpga_lte_fec: not in enabled drivers build config 00:02:22.139 baseband/la12xx: not in enabled drivers build config 00:02:22.139 baseband/null: not in enabled drivers build config 00:02:22.139 baseband/turbo_sw: not in enabled drivers build config 00:02:22.139 gpu/cuda: not in enabled drivers build config 00:02:22.139 00:02:22.139 00:02:22.139 Build targets in project: 311 00:02:22.139 00:02:22.139 DPDK 22.11.4 00:02:22.139 00:02:22.139 User defined options 00:02:22.139 libdir : lib 00:02:22.139 prefix : /home/vagrant/spdk_repo/dpdk/build 00:02:22.139 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:02:22.139 c_link_args : 00:02:22.139 enable_docs : false 00:02:22.139 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:22.139 enable_kmods : false 00:02:22.139 machine : native 00:02:22.139 tests : false 00:02:22.139 00:02:22.139 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:22.139 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:02:22.398 16:40:43 build_native_dpdk -- common/autobuild_common.sh@192 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:02:22.398 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:22.399 [1/740] Generating lib/rte_telemetry_def with a custom command 00:02:22.399 [2/740] Generating lib/rte_kvargs_def with a custom command 00:02:22.399 [3/740] Generating lib/rte_kvargs_mingw with a custom command 00:02:22.399 [4/740] Generating lib/rte_telemetry_mingw with a custom command 00:02:22.399 [5/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:22.399 [6/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:22.399 [7/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:22.399 [8/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:22.399 [9/740] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:22.399 [10/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:22.399 [11/740] Linking static target lib/librte_kvargs.a 00:02:22.399 [12/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:22.399 [13/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:22.399 [14/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:22.658 [15/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:22.658 [16/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:22.658 [17/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:22.658 [18/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:22.658 [19/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:22.658 [20/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:22.658 [21/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_log.c.o 00:02:22.658 [22/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:22.658 [23/740] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:22.658 [24/740] Linking target lib/librte_kvargs.so.23.0 00:02:22.658 [25/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:22.658 [26/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:22.658 [27/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:22.917 [28/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:22.917 [29/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:22.917 [30/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:22.917 [31/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:22.917 [32/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:22.917 [33/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:22.917 [34/740] Linking static target lib/librte_telemetry.a 00:02:22.917 [35/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:22.917 [36/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:22.917 [37/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:22.917 [38/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:22.917 [39/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:22.917 [40/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:22.917 [41/740] Generating symbol file lib/librte_kvargs.so.23.0.p/librte_kvargs.so.23.0.symbols 00:02:23.177 [42/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:23.177 [43/740] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:23.177 [44/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:23.177 [45/740] Linking target lib/librte_telemetry.so.23.0 00:02:23.177 [46/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:23.177 [47/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:23.177 [48/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:23.177 [49/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:23.177 [50/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:23.177 [51/740] Generating symbol file lib/librte_telemetry.so.23.0.p/librte_telemetry.so.23.0.symbols 00:02:23.177 [52/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:23.177 [53/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:23.177 [54/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:23.177 [55/740] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:23.437 [56/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:23.437 [57/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:23.437 [58/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:23.437 [59/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:23.437 [60/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:23.437 [61/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:23.437 [62/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:23.437 [63/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:23.437 [64/740] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:23.437 [65/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:23.437 [66/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_log.c.o 00:02:23.437 [67/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:23.437 [68/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:23.437 [69/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:23.437 [70/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:23.437 [71/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:23.437 [72/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:23.437 [73/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:23.437 [74/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:23.437 [75/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:23.698 [76/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:23.698 [77/740] Generating lib/rte_eal_def with a custom command 00:02:23.698 [78/740] Generating lib/rte_eal_mingw with a custom command 00:02:23.698 [79/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:23.698 [80/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:23.698 [81/740] Generating lib/rte_ring_def with a custom command 00:02:23.698 [82/740] Generating lib/rte_ring_mingw with a custom command 00:02:23.698 [83/740] Generating lib/rte_rcu_def with a custom command 00:02:23.698 [84/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:23.698 [85/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:23.698 [86/740] Generating lib/rte_rcu_mingw with a custom command 00:02:23.698 [87/740] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:23.698 [88/740] Linking static target lib/librte_ring.a 00:02:23.698 [89/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:23.698 [90/740] Generating lib/rte_mempool_def with a custom command 00:02:23.698 [91/740] Generating lib/rte_mempool_mingw with a custom command 00:02:23.698 [92/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:23.958 [93/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:23.958 [94/740] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:23.958 [95/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:23.958 [96/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:23.958 [97/740] Generating lib/rte_mbuf_def with a custom command 00:02:23.958 [98/740] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:23.958 [99/740] Linking static target lib/librte_eal.a 00:02:23.958 [100/740] Generating lib/rte_mbuf_mingw with a custom command 00:02:23.958 [101/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:24.218 [102/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:24.218 [103/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:24.218 [104/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:24.218 [105/740] Linking static target lib/librte_mempool.a 00:02:24.477 [106/740] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:24.477 [107/740] Linking static target lib/librte_rcu.a 00:02:24.477 [108/740] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:24.477 [109/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:24.477 [110/740] Generating lib/rte_net_def with a custom command 00:02:24.477 [111/740] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:24.477 [112/740] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:24.477 [113/740] Generating lib/rte_net_mingw with a custom command 00:02:24.477 [114/740] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:24.477 [115/740] Generating lib/rte_meter_def with a custom command 00:02:24.477 [116/740] Generating lib/rte_meter_mingw with a custom command 00:02:24.477 [117/740] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:24.477 [118/740] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:24.477 [119/740] Linking static target lib/librte_meter.a 00:02:24.737 [120/740] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:24.737 [121/740] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.737 [122/740] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:24.737 [123/740] Linking static target lib/librte_net.a 00:02:24.737 [124/740] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.737 [125/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:24.737 [126/740] Linking static target lib/librte_mbuf.a 00:02:24.997 [127/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:24.997 [128/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:24.997 [129/740] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.997 [130/740] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:24.997 [131/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:24.997 [132/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:24.997 [133/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:25.257 [134/740] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.257 [135/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:25.257 [136/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:25.257 [137/740] Generating lib/rte_ethdev_def with a custom command 00:02:25.519 [138/740] Generating lib/rte_ethdev_mingw with a custom command 00:02:25.519 [139/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:25.519 [140/740] Generating lib/rte_pci_def with a custom command 00:02:25.519 [141/740] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:25.519 [142/740] Linking static target lib/librte_pci.a 00:02:25.519 [143/740] Generating lib/rte_pci_mingw with a custom command 00:02:25.519 [144/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:25.519 [145/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:25.519 [146/740] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:25.519 [147/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:25.519 [148/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:25.779 [149/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:25.779 [150/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:25.779 [151/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:25.779 [152/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:25.779 [153/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:25.779 [154/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:25.779 [155/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:25.779 [156/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:25.779 [157/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:25.779 [158/740] Generating lib/rte_cmdline_def with a custom command 00:02:25.779 [159/740] Generating lib/rte_cmdline_mingw with a custom command 00:02:25.779 [160/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:25.779 [161/740] Generating lib/rte_metrics_def with a custom command 00:02:25.779 [162/740] Generating lib/rte_metrics_mingw with a custom command 00:02:26.039 [163/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:02:26.039 [164/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:26.039 [165/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:26.039 [166/740] Generating lib/rte_hash_def with a custom command 00:02:26.039 [167/740] Generating lib/rte_hash_mingw with a custom command 00:02:26.039 [168/740] Generating lib/rte_timer_def with a custom command 00:02:26.039 [169/740] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:26.039 [170/740] Generating lib/rte_timer_mingw with a custom command 00:02:26.039 [171/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:26.039 [172/740] Linking static target lib/librte_cmdline.a 00:02:26.039 [173/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:26.298 [174/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:02:26.298 [175/740] Linking static target lib/librte_metrics.a 00:02:26.298 [176/740] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:26.298 [177/740] Linking static target lib/librte_timer.a 00:02:26.559 [178/740] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.559 [179/740] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:02:26.559 [180/740] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:26.559 [181/740] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.559 [182/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:02:26.819 [183/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:26.819 [184/740] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:26.819 [185/740] Linking static target lib/librte_ethdev.a 00:02:26.819 [186/740] Generating lib/rte_acl_def with a custom command 00:02:26.819 [187/740] Generating lib/rte_acl_mingw with a custom command 00:02:26.819 [188/740] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:02:26.819 [189/740] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:02:26.819 [190/740] Generating lib/rte_bbdev_def with a custom command 00:02:26.819 [191/740] Generating lib/rte_bbdev_mingw with a custom command 00:02:26.819 [192/740] Generating lib/rte_bitratestats_def with a custom command 00:02:27.079 [193/740] Generating lib/rte_bitratestats_mingw with a custom command 00:02:27.339 [194/740] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:02:27.339 [195/740] Linking static target lib/librte_bitratestats.a 00:02:27.339 [196/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:02:27.339 [197/740] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:27.339 [198/740] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:02:27.598 [199/740] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:02:27.598 [200/740] Linking static target lib/librte_bbdev.a 00:02:27.598 [201/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:02:27.857 [202/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:02:27.857 [203/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:02:28.118 [204/740] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.118 [205/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:02:28.118 [206/740] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:28.118 [207/740] Linking static target lib/librte_hash.a 00:02:28.118 [208/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:02:28.378 [209/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:02:28.378 [210/740] Generating lib/rte_bpf_def with a custom command 00:02:28.378 [211/740] Generating lib/rte_bpf_mingw with a custom command 00:02:28.638 [212/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:02:28.638 [213/740] Generating lib/rte_cfgfile_def with a custom command 00:02:28.638 [214/740] Generating lib/rte_cfgfile_mingw with a custom command 00:02:28.638 [215/740] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:02:28.638 [216/740] Linking static target lib/librte_cfgfile.a 00:02:28.638 [217/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:02:28.638 [218/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx2.c.o 00:02:28.638 [219/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:02:28.638 [220/740] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.638 [221/740] Generating lib/rte_compressdev_def with a custom command 00:02:28.638 [222/740] Generating lib/rte_compressdev_mingw with a custom command 00:02:28.898 [223/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:02:28.898 [224/740] Linking static target lib/librte_bpf.a 00:02:28.898 [225/740] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:02:28.898 [226/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:28.898 [227/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:02:28.898 [228/740] Generating lib/rte_cryptodev_def with a custom command 00:02:29.158 [229/740] Linking static target lib/librte_acl.a 00:02:29.158 [230/740] Generating lib/rte_cryptodev_mingw with a custom command 00:02:29.158 [231/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:29.158 [232/740] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.158 [233/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:29.158 [234/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:29.158 [235/740] Linking static target lib/librte_compressdev.a 00:02:29.158 [236/740] Generating lib/rte_distributor_def with a custom command 00:02:29.158 [237/740] Generating lib/rte_distributor_mingw with a custom command 00:02:29.158 [238/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:29.418 [239/740] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.418 [240/740] Generating lib/rte_efd_def with a custom command 00:02:29.418 [241/740] Generating lib/rte_efd_mingw with a custom command 00:02:29.418 [242/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:02:29.418 [243/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:02:29.678 [244/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:02:29.678 [245/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:02:29.678 [246/740] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.678 [247/740] Linking static target lib/librte_distributor.a 00:02:29.678 [248/740] Linking target lib/librte_eal.so.23.0 00:02:29.678 [249/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:02:29.938 [250/740] Generating symbol file lib/librte_eal.so.23.0.p/librte_eal.so.23.0.symbols 00:02:29.938 [251/740] Linking target lib/librte_ring.so.23.0 00:02:29.938 [252/740] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.938 [253/740] Linking target lib/librte_meter.so.23.0 00:02:29.938 [254/740] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:29.938 [255/740] Generating symbol file lib/librte_ring.so.23.0.p/librte_ring.so.23.0.symbols 00:02:29.938 [256/740] Linking target lib/librte_pci.so.23.0 00:02:29.938 [257/740] Linking target lib/librte_rcu.so.23.0 00:02:29.938 [258/740] Generating symbol file lib/librte_meter.so.23.0.p/librte_meter.so.23.0.symbols 00:02:29.938 [259/740] Linking target lib/librte_mempool.so.23.0 00:02:30.198 [260/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:02:30.198 [261/740] Generating symbol file lib/librte_rcu.so.23.0.p/librte_rcu.so.23.0.symbols 00:02:30.198 [262/740] Generating symbol file lib/librte_pci.so.23.0.p/librte_pci.so.23.0.symbols 00:02:30.198 [263/740] Linking target lib/librte_timer.so.23.0 00:02:30.198 [264/740] Generating symbol file lib/librte_mempool.so.23.0.p/librte_mempool.so.23.0.symbols 00:02:30.198 [265/740] Linking target lib/librte_cfgfile.so.23.0 00:02:30.198 [266/740] Linking target lib/librte_acl.so.23.0 00:02:30.198 [267/740] Linking target lib/librte_mbuf.so.23.0 00:02:30.198 [268/740] Generating symbol file lib/librte_timer.so.23.0.p/librte_timer.so.23.0.symbols 00:02:30.198 [269/740] Generating symbol file lib/librte_mbuf.so.23.0.p/librte_mbuf.so.23.0.symbols 00:02:30.198 [270/740] Generating symbol file lib/librte_acl.so.23.0.p/librte_acl.so.23.0.symbols 00:02:30.198 [271/740] Linking target lib/librte_bbdev.so.23.0 00:02:30.198 [272/740] Linking target lib/librte_net.so.23.0 00:02:30.198 [273/740] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:02:30.198 [274/740] Linking target lib/librte_compressdev.so.23.0 00:02:30.198 [275/740] Linking target lib/librte_distributor.so.23.0 00:02:30.198 [276/740] Linking static target lib/librte_efd.a 00:02:30.458 [277/740] Generating lib/rte_eventdev_def with a custom command 00:02:30.458 [278/740] Generating lib/rte_eventdev_mingw with a custom command 00:02:30.458 [279/740] Generating symbol file lib/librte_net.so.23.0.p/librte_net.so.23.0.symbols 00:02:30.458 [280/740] Linking target lib/librte_cmdline.so.23.0 00:02:30.458 [281/740] Linking target lib/librte_hash.so.23.0 00:02:30.459 [282/740] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.459 [283/740] Generating lib/rte_gpudev_def with a custom command 00:02:30.459 [284/740] Generating lib/rte_gpudev_mingw with a custom command 00:02:30.459 [285/740] Generating symbol file lib/librte_hash.so.23.0.p/librte_hash.so.23.0.symbols 00:02:30.459 [286/740] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:30.459 [287/740] Linking target lib/librte_efd.so.23.0 00:02:30.719 [288/740] Linking target lib/librte_ethdev.so.23.0 00:02:30.719 [289/740] Generating symbol file lib/librte_ethdev.so.23.0.p/librte_ethdev.so.23.0.symbols 00:02:30.719 [290/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:02:30.719 [291/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:02:30.719 [292/740] Linking target lib/librte_metrics.so.23.0 00:02:30.719 [293/740] Linking target lib/librte_bpf.so.23.0 00:02:30.719 [294/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:30.719 [295/740] Linking static target lib/librte_cryptodev.a 00:02:30.979 [296/740] Generating symbol file lib/librte_metrics.so.23.0.p/librte_metrics.so.23.0.symbols 00:02:30.979 [297/740] Generating symbol file lib/librte_bpf.so.23.0.p/librte_bpf.so.23.0.symbols 00:02:30.979 [298/740] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:02:30.979 [299/740] Linking static target lib/librte_gpudev.a 00:02:30.979 [300/740] Linking target lib/librte_bitratestats.so.23.0 00:02:30.979 [301/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:02:30.979 [302/740] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:02:30.979 [303/740] Generating lib/rte_gro_def with a custom command 00:02:30.979 [304/740] Generating lib/rte_gro_mingw with a custom command 00:02:30.979 [305/740] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:02:31.257 [306/740] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:02:31.257 [307/740] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:02:31.537 [308/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:02:31.537 [309/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:02:31.537 [310/740] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:02:31.537 [311/740] Generating lib/rte_gso_def with a custom command 00:02:31.537 [312/740] Generating lib/rte_gso_mingw with a custom command 00:02:31.537 [313/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:02:31.537 [314/740] Linking static target lib/librte_gro.a 00:02:31.537 [315/740] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.537 [316/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:02:31.537 [317/740] Linking target lib/librte_gpudev.so.23.0 00:02:31.797 [318/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:02:31.797 [319/740] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:02:31.797 [320/740] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:02:31.797 [321/740] Linking target lib/librte_gro.so.23.0 00:02:31.797 [322/740] Generating lib/rte_ip_frag_def with a custom command 00:02:31.797 [323/740] Generating lib/rte_ip_frag_mingw with a custom command 00:02:31.797 [324/740] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:02:31.797 [325/740] Linking static target lib/librte_gso.a 00:02:31.797 [326/740] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:02:32.057 [327/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:02:32.057 [328/740] Linking static target lib/librte_jobstats.a 00:02:32.057 [329/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:02:32.057 [330/740] Linking static target lib/librte_eventdev.a 00:02:32.057 [331/740] Generating lib/rte_jobstats_def with a custom command 00:02:32.057 [332/740] Generating lib/rte_jobstats_mingw with a custom command 00:02:32.057 [333/740] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.057 [334/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:02:32.057 [335/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:02:32.057 [336/740] Generating lib/rte_latencystats_mingw with a custom command 00:02:32.057 [337/740] Generating lib/rte_latencystats_def with a custom command 00:02:32.057 [338/740] Linking target lib/librte_gso.so.23.0 00:02:32.057 [339/740] Generating lib/rte_lpm_def with a custom command 00:02:32.057 [340/740] Generating lib/rte_lpm_mingw with a custom command 00:02:32.057 [341/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:02:32.057 [342/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:02:32.317 [343/740] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.317 [344/740] Linking target lib/librte_jobstats.so.23.0 00:02:32.317 [345/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:02:32.317 [346/740] Linking static target lib/librte_ip_frag.a 00:02:32.578 [347/740] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:02:32.578 [348/740] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:02:32.578 [349/740] Linking static target lib/librte_latencystats.a 00:02:32.578 [350/740] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:02:32.578 [351/740] Linking static target lib/member/libsketch_avx512_tmp.a 00:02:32.578 [352/740] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.578 [353/740] Generating lib/rte_member_def with a custom command 00:02:32.578 [354/740] Linking target lib/librte_ip_frag.so.23.0 00:02:32.578 [355/740] Generating lib/rte_member_mingw with a custom command 00:02:32.578 [356/740] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.578 [357/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:02:32.578 [358/740] Linking target lib/librte_cryptodev.so.23.0 00:02:32.578 [359/740] Generating lib/rte_pcapng_def with a custom command 00:02:32.578 [360/740] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:32.578 [361/740] Generating symbol file lib/librte_ip_frag.so.23.0.p/librte_ip_frag.so.23.0.symbols 00:02:32.578 [362/740] Generating lib/rte_pcapng_mingw with a custom command 00:02:32.578 [363/740] Linking target lib/librte_latencystats.so.23.0 00:02:32.838 [364/740] Generating symbol file lib/librte_cryptodev.so.23.0.p/librte_cryptodev.so.23.0.symbols 00:02:32.838 [365/740] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:32.838 [366/740] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:32.838 [367/740] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:32.838 [368/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:02:32.838 [369/740] Linking static target lib/librte_lpm.a 00:02:32.838 [370/740] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:33.098 [371/740] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:02:33.098 [372/740] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:02:33.098 [373/740] Compiling C object lib/librte_power.a.p/power_rte_power_empty_poll.c.o 00:02:33.098 [374/740] Generating lib/rte_power_def with a custom command 00:02:33.098 [375/740] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:33.098 [376/740] Generating lib/rte_power_mingw with a custom command 00:02:33.098 [377/740] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.098 [378/740] Generating lib/rte_rawdev_def with a custom command 00:02:33.098 [379/740] Linking target lib/librte_lpm.so.23.0 00:02:33.098 [380/740] Generating lib/rte_rawdev_mingw with a custom command 00:02:33.098 [381/740] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:33.098 [382/740] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:02:33.358 [383/740] Linking static target lib/librte_pcapng.a 00:02:33.358 [384/740] Generating lib/rte_regexdev_def with a custom command 00:02:33.358 [385/740] Generating lib/rte_regexdev_mingw with a custom command 00:02:33.358 [386/740] Generating symbol file lib/librte_lpm.so.23.0.p/librte_lpm.so.23.0.symbols 00:02:33.358 [387/740] Generating lib/rte_dmadev_def with a custom command 00:02:33.358 [388/740] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:33.358 [389/740] Generating lib/rte_dmadev_mingw with a custom command 00:02:33.358 [390/740] Compiling C object lib/librte_power.a.p/power_rte_power_intel_uncore.c.o 00:02:33.358 [391/740] Generating lib/rte_rib_def with a custom command 00:02:33.358 [392/740] Generating lib/rte_rib_mingw with a custom command 00:02:33.358 [393/740] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:02:33.358 [394/740] Linking static target lib/librte_rawdev.a 00:02:33.358 [395/740] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:33.358 [396/740] Linking static target lib/librte_power.a 00:02:33.358 [397/740] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.618 [398/740] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.618 [399/740] Linking target lib/librte_pcapng.so.23.0 00:02:33.618 [400/740] Linking target lib/librte_eventdev.so.23.0 00:02:33.618 [401/740] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:33.618 [402/740] Linking static target lib/librte_dmadev.a 00:02:33.618 [403/740] Generating symbol file lib/librte_pcapng.so.23.0.p/librte_pcapng.so.23.0.symbols 00:02:33.618 [404/740] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:02:33.618 [405/740] Linking static target lib/librte_regexdev.a 00:02:33.618 [406/740] Generating lib/rte_reorder_def with a custom command 00:02:33.618 [407/740] Generating lib/rte_reorder_mingw with a custom command 00:02:33.618 [408/740] Generating symbol file lib/librte_eventdev.so.23.0.p/librte_eventdev.so.23.0.symbols 00:02:33.878 [409/740] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:02:33.878 [410/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:02:33.878 [411/740] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:33.878 [412/740] Linking target lib/librte_rawdev.so.23.0 00:02:33.878 [413/740] Generating lib/rte_sched_def with a custom command 00:02:33.878 [414/740] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:02:33.878 [415/740] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:02:33.878 [416/740] Generating lib/rte_sched_mingw with a custom command 00:02:33.878 [417/740] Generating lib/rte_security_def with a custom command 00:02:33.878 [418/740] Generating lib/rte_security_mingw with a custom command 00:02:33.878 [419/740] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:02:33.878 [420/740] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:33.878 [421/740] Linking static target lib/librte_reorder.a 00:02:33.878 [422/740] Linking static target lib/librte_member.a 00:02:34.137 [423/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:02:34.137 [424/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:02:34.137 [425/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:02:34.137 [426/740] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.137 [427/740] Linking static target lib/librte_rib.a 00:02:34.137 [428/740] Generating lib/rte_stack_def with a custom command 00:02:34.137 [429/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:02:34.137 [430/740] Linking target lib/librte_dmadev.so.23.0 00:02:34.137 [431/740] Generating lib/rte_stack_mingw with a custom command 00:02:34.137 [432/740] Linking static target lib/librte_stack.a 00:02:34.137 [433/740] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.137 [434/740] Generating symbol file lib/librte_dmadev.so.23.0.p/librte_dmadev.so.23.0.symbols 00:02:34.138 [435/740] Linking target lib/librte_reorder.so.23.0 00:02:34.138 [436/740] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:34.138 [437/740] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.138 [438/740] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.138 [439/740] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.138 [440/740] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.138 [441/740] Linking target lib/librte_regexdev.so.23.0 00:02:34.397 [442/740] Linking target lib/librte_power.so.23.0 00:02:34.397 [443/740] Linking target lib/librte_member.so.23.0 00:02:34.397 [444/740] Linking target lib/librte_stack.so.23.0 00:02:34.397 [445/740] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:34.397 [446/740] Linking static target lib/librte_security.a 00:02:34.397 [447/740] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.397 [448/740] Linking target lib/librte_rib.so.23.0 00:02:34.656 [449/740] Generating symbol file lib/librte_rib.so.23.0.p/librte_rib.so.23.0.symbols 00:02:34.656 [450/740] Generating lib/rte_vhost_def with a custom command 00:02:34.656 [451/740] Generating lib/rte_vhost_mingw with a custom command 00:02:34.656 [452/740] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:34.656 [453/740] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:34.656 [454/740] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:34.656 [455/740] Linking target lib/librte_security.so.23.0 00:02:34.915 [456/740] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:34.915 [457/740] Generating symbol file lib/librte_security.so.23.0.p/librte_security.so.23.0.symbols 00:02:35.175 [458/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:02:35.175 [459/740] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:02:35.175 [460/740] Linking static target lib/librte_sched.a 00:02:35.175 [461/740] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:02:35.175 [462/740] Generating lib/rte_ipsec_def with a custom command 00:02:35.175 [463/740] Generating lib/rte_ipsec_mingw with a custom command 00:02:35.435 [464/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:35.435 [465/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:35.435 [466/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:02:35.435 [467/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:02:35.435 [468/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:02:35.435 [469/740] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:02:35.435 [470/740] Linking target lib/librte_sched.so.23.0 00:02:35.696 [471/740] Generating symbol file lib/librte_sched.so.23.0.p/librte_sched.so.23.0.symbols 00:02:35.696 [472/740] Generating lib/rte_fib_def with a custom command 00:02:35.696 [473/740] Generating lib/rte_fib_mingw with a custom command 00:02:35.696 [474/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:02:35.955 [475/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:02:35.955 [476/740] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:02:35.955 [477/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:02:36.215 [478/740] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:02:36.215 [479/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:02:36.215 [480/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:02:36.215 [481/740] Linking static target lib/librte_ipsec.a 00:02:36.215 [482/740] Linking static target lib/librte_fib.a 00:02:36.475 [483/740] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:02:36.475 [484/740] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:02:36.475 [485/740] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:02:36.475 [486/740] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:02:36.475 [487/740] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:02:36.475 [488/740] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.475 [489/740] Linking target lib/librte_fib.so.23.0 00:02:36.475 [490/740] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:02:36.475 [491/740] Linking target lib/librte_ipsec.so.23.0 00:02:37.043 [492/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:02:37.043 [493/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:02:37.043 [494/740] Generating lib/rte_port_def with a custom command 00:02:37.043 [495/740] Generating lib/rte_port_mingw with a custom command 00:02:37.043 [496/740] Generating lib/rte_pdump_def with a custom command 00:02:37.043 [497/740] Generating lib/rte_pdump_mingw with a custom command 00:02:37.043 [498/740] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:02:37.043 [499/740] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:02:37.043 [500/740] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:02:37.043 [501/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:02:37.302 [502/740] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:02:37.302 [503/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:02:37.302 [504/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:02:37.302 [505/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:02:37.560 [506/740] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:02:37.560 [507/740] Linking static target lib/librte_port.a 00:02:37.560 [508/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:02:37.560 [509/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:02:37.818 [510/740] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:02:37.818 [511/740] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:02:37.818 [512/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:02:37.818 [513/740] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:02:37.818 [514/740] Linking static target lib/librte_pdump.a 00:02:37.818 [515/740] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:02:38.077 [516/740] Linking target lib/librte_port.so.23.0 00:02:38.077 [517/740] Generating symbol file lib/librte_port.so.23.0.p/librte_port.so.23.0.symbols 00:02:38.077 [518/740] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:02:38.077 [519/740] Linking target lib/librte_pdump.so.23.0 00:02:38.335 [520/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:02:38.335 [521/740] Generating lib/rte_table_def with a custom command 00:02:38.335 [522/740] Generating lib/rte_table_mingw with a custom command 00:02:38.335 [523/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:02:38.335 [524/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:02:38.335 [525/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:02:38.593 [526/740] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:02:38.593 [527/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:02:38.593 [528/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:02:38.593 [529/740] Generating lib/rte_pipeline_def with a custom command 00:02:38.593 [530/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:02:38.593 [531/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:02:38.593 [532/740] Generating lib/rte_pipeline_mingw with a custom command 00:02:38.593 [533/740] Linking static target lib/librte_table.a 00:02:38.852 [534/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:02:39.111 [535/740] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:02:39.111 [536/740] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:02:39.111 [537/740] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:02:39.111 [538/740] Linking target lib/librte_table.so.23.0 00:02:39.370 [539/740] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:02:39.370 [540/740] Generating symbol file lib/librte_table.so.23.0.p/librte_table.so.23.0.symbols 00:02:39.370 [541/740] Generating lib/rte_graph_def with a custom command 00:02:39.370 [542/740] Generating lib/rte_graph_mingw with a custom command 00:02:39.370 [543/740] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:02:39.629 [544/740] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:02:39.629 [545/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:02:39.629 [546/740] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:02:39.629 [547/740] Linking static target lib/librte_graph.a 00:02:39.629 [548/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:39.629 [549/740] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:02:39.887 [550/740] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:02:39.887 [551/740] Compiling C object lib/librte_node.a.p/node_null.c.o 00:02:39.887 [552/740] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:02:40.146 [553/740] Compiling C object lib/librte_node.a.p/node_log.c.o 00:02:40.146 [554/740] Generating lib/rte_node_def with a custom command 00:02:40.146 [555/740] Generating lib/rte_node_mingw with a custom command 00:02:40.146 [556/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:02:40.146 [557/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:40.146 [558/740] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:02:40.146 [559/740] Linking target lib/librte_graph.so.23.0 00:02:40.405 [560/740] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:02:40.405 [561/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:40.405 [562/740] Generating symbol file lib/librte_graph.so.23.0.p/librte_graph.so.23.0.symbols 00:02:40.405 [563/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:40.405 [564/740] Generating drivers/rte_bus_pci_def with a custom command 00:02:40.405 [565/740] Generating drivers/rte_bus_pci_mingw with a custom command 00:02:40.405 [566/740] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:02:40.405 [567/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:40.405 [568/740] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:02:40.405 [569/740] Generating drivers/rte_bus_vdev_def with a custom command 00:02:40.663 [570/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:40.663 [571/740] Generating drivers/rte_bus_vdev_mingw with a custom command 00:02:40.663 [572/740] Generating drivers/rte_mempool_ring_def with a custom command 00:02:40.663 [573/740] Generating drivers/rte_mempool_ring_mingw with a custom command 00:02:40.663 [574/740] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:02:40.663 [575/740] Linking static target lib/librte_node.a 00:02:40.663 [576/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:40.663 [577/740] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:40.663 [578/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:40.663 [579/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:40.663 [580/740] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:40.923 [581/740] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:02:40.923 [582/740] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:40.923 [583/740] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:40.923 [584/740] Linking static target drivers/librte_bus_vdev.a 00:02:40.923 [585/740] Linking target lib/librte_node.so.23.0 00:02:40.923 [586/740] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:40.923 [587/740] Compiling C object drivers/librte_bus_vdev.so.23.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:40.923 [588/740] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:40.923 [589/740] Compiling C object drivers/librte_bus_pci.so.23.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:40.923 [590/740] Linking static target drivers/librte_bus_pci.a 00:02:40.923 [591/740] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.182 [592/740] Linking target drivers/librte_bus_vdev.so.23.0 00:02:41.182 [593/740] Generating symbol file drivers/librte_bus_vdev.so.23.0.p/librte_bus_vdev.so.23.0.symbols 00:02:41.182 [594/740] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.182 [595/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:02:41.182 [596/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:02:41.182 [597/740] Linking target drivers/librte_bus_pci.so.23.0 00:02:41.182 [598/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:02:41.442 [599/740] Generating symbol file drivers/librte_bus_pci.so.23.0.p/librte_bus_pci.so.23.0.symbols 00:02:41.442 [600/740] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:41.442 [601/740] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:41.442 [602/740] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:41.442 [603/740] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:41.442 [604/740] Linking static target drivers/librte_mempool_ring.a 00:02:41.442 [605/740] Compiling C object drivers/librte_mempool_ring.so.23.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:41.442 [606/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:02:41.442 [607/740] Linking target drivers/librte_mempool_ring.so.23.0 00:02:41.702 [608/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:02:41.961 [609/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:02:42.221 [610/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:02:42.221 [611/740] Linking static target drivers/net/i40e/base/libi40e_base.a 00:02:42.221 [612/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:02:42.791 [613/740] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:02:42.791 [614/740] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:02:42.791 [615/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:02:42.791 [616/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:02:43.050 [617/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:02:43.050 [618/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:02:43.050 [619/740] Generating drivers/rte_net_i40e_def with a custom command 00:02:43.050 [620/740] Generating drivers/rte_net_i40e_mingw with a custom command 00:02:43.310 [621/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:02:43.310 [622/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:02:43.879 [623/740] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:02:44.137 [624/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:02:44.137 [625/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:02:44.137 [626/740] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:02:44.138 [627/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:02:44.138 [628/740] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:02:44.396 [629/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:02:44.396 [630/740] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:02:44.396 [631/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:02:44.396 [632/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:02:44.396 [633/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_avx2.c.o 00:02:44.963 [634/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:02:44.963 [635/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:02:44.963 [636/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:02:44.963 [637/740] Linking static target drivers/libtmp_rte_net_i40e.a 00:02:44.963 [638/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:02:44.963 [639/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:02:45.223 [640/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:02:45.223 [641/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:02:45.223 [642/740] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:02:45.223 [643/740] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:45.223 [644/740] Linking static target drivers/librte_net_i40e.a 00:02:45.483 [645/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:02:45.483 [646/740] Compiling C object drivers/librte_net_i40e.so.23.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:02:45.483 [647/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:02:45.743 [648/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:02:45.743 [649/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:02:45.743 [650/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:02:45.743 [651/740] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.743 [652/740] Linking target drivers/librte_net_i40e.so.23.0 00:02:46.002 [653/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:02:46.002 [654/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:02:46.002 [655/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:02:46.002 [656/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:02:46.002 [657/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:02:46.262 [658/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:02:46.262 [659/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:02:46.262 [660/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:02:46.262 [661/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:02:46.262 [662/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:02:46.522 [663/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:02:46.522 [664/740] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:02:46.522 [665/740] Linking static target lib/librte_vhost.a 00:02:46.784 [666/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:02:46.784 [667/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:02:47.044 [668/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:02:47.304 [669/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:02:47.304 [670/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:02:47.304 [671/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:02:47.565 [672/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:02:47.565 [673/740] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.565 [674/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:02:47.565 [675/740] Linking target lib/librte_vhost.so.23.0 00:02:47.565 [676/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:02:47.825 [677/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:02:47.825 [678/740] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:02:47.825 [679/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:02:47.825 [680/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:02:47.825 [681/740] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:02:48.085 [682/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:02:48.085 [683/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:02:48.085 [684/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:02:48.085 [685/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:02:48.345 [686/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:02:48.345 [687/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:02:48.345 [688/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:02:48.345 [689/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:02:48.345 [690/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:02:48.605 [691/740] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:02:48.865 [692/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:02:48.865 [693/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:02:48.865 [694/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:02:48.865 [695/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:02:48.865 [696/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:02:49.435 [697/740] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:02:49.435 [698/740] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:02:49.435 [699/740] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:02:49.435 [700/740] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:02:49.435 [701/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:02:50.006 [702/740] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:02:50.006 [703/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:02:50.006 [704/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:02:50.006 [705/740] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:02:50.006 [706/740] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:02:50.266 [707/740] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:02:50.266 [708/740] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:02:50.527 [709/740] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:02:50.787 [710/740] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:02:50.787 [711/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:02:50.787 [712/740] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:02:50.787 [713/740] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:02:51.047 [714/740] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:02:51.047 [715/740] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:02:51.047 [716/740] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:02:51.307 [717/740] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:02:51.307 [718/740] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:02:51.567 [719/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:02:51.567 [720/740] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:02:51.567 [721/740] Linking static target lib/librte_pipeline.a 00:02:51.826 [722/740] Linking target app/dpdk-test-cmdline 00:02:52.085 [723/740] Linking target app/dpdk-proc-info 00:02:52.085 [724/740] Linking target app/dpdk-test-eventdev 00:02:52.085 [725/740] Linking target app/dpdk-test-bbdev 00:02:52.085 [726/740] Linking target app/dpdk-dumpcap 00:02:52.085 [727/740] Linking target app/dpdk-test-acl 00:02:52.085 [728/740] Linking target app/dpdk-pdump 00:02:52.085 [729/740] Linking target app/dpdk-test-compress-perf 00:02:52.085 [730/740] Linking target app/dpdk-test-crypto-perf 00:02:52.344 [731/740] Linking target app/dpdk-test-gpudev 00:02:52.344 [732/740] Linking target app/dpdk-test-flow-perf 00:02:52.344 [733/740] Linking target app/dpdk-test-fib 00:02:52.344 [734/740] Linking target app/dpdk-test-pipeline 00:02:52.344 [735/740] Linking target app/dpdk-testpmd 00:02:52.344 [736/740] Linking target app/dpdk-test-sad 00:02:52.344 [737/740] Linking target app/dpdk-test-security-perf 00:02:52.345 [738/740] Linking target app/dpdk-test-regex 00:02:56.546 [739/740] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.805 [740/740] Linking target lib/librte_pipeline.so.23.0 00:02:56.805 16:41:18 build_native_dpdk -- common/autobuild_common.sh@194 -- $ uname -s 00:02:56.805 16:41:18 build_native_dpdk -- common/autobuild_common.sh@194 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:02:56.805 16:41:18 build_native_dpdk -- common/autobuild_common.sh@207 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:02:56.805 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:56.805 [0/1] Installing files. 00:02:57.067 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.067 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/flow_classify.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/ipv4_rules_file.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/kni.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.068 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.069 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.070 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.332 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:57.333 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:57.334 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:02:57.334 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.334 Installing lib/librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing lib/librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing drivers/librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:57.335 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing drivers/librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:57.335 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing drivers/librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:57.335 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:02:57.335 Installing drivers/librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:02:57.335 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.335 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.597 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.598 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_empty_poll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_intel_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.599 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:57.600 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:02:57.600 Installing symlink pointing to librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.23 00:02:57.600 Installing symlink pointing to librte_kvargs.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:02:57.600 Installing symlink pointing to librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.23 00:02:57.600 Installing symlink pointing to librte_telemetry.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:02:57.600 Installing symlink pointing to librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.23 00:02:57.600 Installing symlink pointing to librte_eal.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:02:57.600 Installing symlink pointing to librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.23 00:02:57.600 Installing symlink pointing to librte_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:02:57.600 Installing symlink pointing to librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.23 00:02:57.600 Installing symlink pointing to librte_rcu.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:02:57.600 Installing symlink pointing to librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.23 00:02:57.600 Installing symlink pointing to librte_mempool.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:02:57.600 Installing symlink pointing to librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.23 00:02:57.600 Installing symlink pointing to librte_mbuf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:02:57.600 Installing symlink pointing to librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.23 00:02:57.600 Installing symlink pointing to librte_net.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:02:57.600 Installing symlink pointing to librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.23 00:02:57.600 Installing symlink pointing to librte_meter.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:02:57.600 Installing symlink pointing to librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.23 00:02:57.600 Installing symlink pointing to librte_ethdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:02:57.600 Installing symlink pointing to librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.23 00:02:57.600 Installing symlink pointing to librte_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:02:57.600 Installing symlink pointing to librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.23 00:02:57.600 Installing symlink pointing to librte_cmdline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:02:57.600 Installing symlink pointing to librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.23 00:02:57.600 Installing symlink pointing to librte_metrics.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:02:57.600 Installing symlink pointing to librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.23 00:02:57.601 Installing symlink pointing to librte_hash.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:02:57.601 Installing symlink pointing to librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.23 00:02:57.601 Installing symlink pointing to librte_timer.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:02:57.601 Installing symlink pointing to librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.23 00:02:57.601 Installing symlink pointing to librte_acl.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:02:57.601 Installing symlink pointing to librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.23 00:02:57.601 Installing symlink pointing to librte_bbdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:02:57.601 Installing symlink pointing to librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.23 00:02:57.601 Installing symlink pointing to librte_bitratestats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:02:57.601 Installing symlink pointing to librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.23 00:02:57.601 Installing symlink pointing to librte_bpf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:02:57.601 Installing symlink pointing to librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.23 00:02:57.601 Installing symlink pointing to librte_cfgfile.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:02:57.601 Installing symlink pointing to librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.23 00:02:57.601 Installing symlink pointing to librte_compressdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:02:57.601 Installing symlink pointing to librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.23 00:02:57.601 Installing symlink pointing to librte_cryptodev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:02:57.601 Installing symlink pointing to librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.23 00:02:57.601 Installing symlink pointing to librte_distributor.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:02:57.601 Installing symlink pointing to librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.23 00:02:57.601 Installing symlink pointing to librte_efd.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:02:57.601 Installing symlink pointing to librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.23 00:02:57.601 Installing symlink pointing to librte_eventdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:02:57.601 Installing symlink pointing to librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.23 00:02:57.601 './librte_bus_pci.so' -> 'dpdk/pmds-23.0/librte_bus_pci.so' 00:02:57.601 './librte_bus_pci.so.23' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23' 00:02:57.601 './librte_bus_pci.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23.0' 00:02:57.601 './librte_bus_vdev.so' -> 'dpdk/pmds-23.0/librte_bus_vdev.so' 00:02:57.601 './librte_bus_vdev.so.23' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23' 00:02:57.601 './librte_bus_vdev.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23.0' 00:02:57.601 './librte_mempool_ring.so' -> 'dpdk/pmds-23.0/librte_mempool_ring.so' 00:02:57.601 './librte_mempool_ring.so.23' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23' 00:02:57.601 './librte_mempool_ring.so.23.0' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23.0' 00:02:57.601 './librte_net_i40e.so' -> 'dpdk/pmds-23.0/librte_net_i40e.so' 00:02:57.601 './librte_net_i40e.so.23' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23' 00:02:57.601 './librte_net_i40e.so.23.0' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23.0' 00:02:57.601 Installing symlink pointing to librte_gpudev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:02:57.601 Installing symlink pointing to librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.23 00:02:57.601 Installing symlink pointing to librte_gro.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:02:57.601 Installing symlink pointing to librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.23 00:02:57.601 Installing symlink pointing to librte_gso.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:02:57.601 Installing symlink pointing to librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.23 00:02:57.601 Installing symlink pointing to librte_ip_frag.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:02:57.601 Installing symlink pointing to librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.23 00:02:57.601 Installing symlink pointing to librte_jobstats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:02:57.601 Installing symlink pointing to librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.23 00:02:57.601 Installing symlink pointing to librte_latencystats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:02:57.601 Installing symlink pointing to librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.23 00:02:57.601 Installing symlink pointing to librte_lpm.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:02:57.601 Installing symlink pointing to librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.23 00:02:57.601 Installing symlink pointing to librte_member.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:02:57.601 Installing symlink pointing to librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.23 00:02:57.601 Installing symlink pointing to librte_pcapng.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:02:57.601 Installing symlink pointing to librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.23 00:02:57.601 Installing symlink pointing to librte_power.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:02:57.601 Installing symlink pointing to librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.23 00:02:57.601 Installing symlink pointing to librte_rawdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:02:57.601 Installing symlink pointing to librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.23 00:02:57.601 Installing symlink pointing to librte_regexdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:02:57.601 Installing symlink pointing to librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.23 00:02:57.601 Installing symlink pointing to librte_dmadev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:02:57.601 Installing symlink pointing to librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.23 00:02:57.601 Installing symlink pointing to librte_rib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:02:57.601 Installing symlink pointing to librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.23 00:02:57.601 Installing symlink pointing to librte_reorder.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:02:57.601 Installing symlink pointing to librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.23 00:02:57.601 Installing symlink pointing to librte_sched.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:02:57.601 Installing symlink pointing to librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.23 00:02:57.601 Installing symlink pointing to librte_security.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:02:57.601 Installing symlink pointing to librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.23 00:02:57.601 Installing symlink pointing to librte_stack.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:02:57.601 Installing symlink pointing to librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.23 00:02:57.601 Installing symlink pointing to librte_vhost.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:02:57.601 Installing symlink pointing to librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.23 00:02:57.601 Installing symlink pointing to librte_ipsec.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:02:57.601 Installing symlink pointing to librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.23 00:02:57.601 Installing symlink pointing to librte_fib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:02:57.601 Installing symlink pointing to librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.23 00:02:57.601 Installing symlink pointing to librte_port.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:02:57.601 Installing symlink pointing to librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.23 00:02:57.601 Installing symlink pointing to librte_pdump.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:02:57.601 Installing symlink pointing to librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.23 00:02:57.601 Installing symlink pointing to librte_table.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:02:57.601 Installing symlink pointing to librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.23 00:02:57.601 Installing symlink pointing to librte_pipeline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:02:57.601 Installing symlink pointing to librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.23 00:02:57.601 Installing symlink pointing to librte_graph.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:02:57.601 Installing symlink pointing to librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.23 00:02:57.601 Installing symlink pointing to librte_node.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:02:57.601 Installing symlink pointing to librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23 00:02:57.601 Installing symlink pointing to librte_bus_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:02:57.601 Installing symlink pointing to librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23 00:02:57.601 Installing symlink pointing to librte_bus_vdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:02:57.601 Installing symlink pointing to librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23 00:02:57.601 Installing symlink pointing to librte_mempool_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:02:57.601 Installing symlink pointing to librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23 00:02:57.601 Installing symlink pointing to librte_net_i40e.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:02:57.601 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-23.0' 00:02:57.601 16:41:19 build_native_dpdk -- common/autobuild_common.sh@213 -- $ cat 00:02:57.601 16:41:19 build_native_dpdk -- common/autobuild_common.sh@218 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:57.601 00:02:57.601 real 0m42.421s 00:02:57.601 user 4m7.879s 00:02:57.601 sys 0m50.328s 00:02:57.601 16:41:19 build_native_dpdk -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:57.601 ************************************ 00:02:57.601 END TEST build_native_dpdk 00:02:57.601 ************************************ 00:02:57.601 16:41:19 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:02:57.601 16:41:19 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:02:57.601 16:41:19 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:02:57.601 16:41:19 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:02:57.601 16:41:19 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:02:57.601 16:41:19 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:02:57.601 16:41:19 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:02:57.601 16:41:19 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:02:57.602 16:41:19 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:02:57.861 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:02:58.121 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:02:58.121 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:02:58.121 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:02:58.423 Using 'verbs' RDMA provider 00:03:14.258 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:03:29.154 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:03:29.725 Creating mk/config.mk...done. 00:03:29.725 Creating mk/cc.flags.mk...done. 00:03:29.725 Type 'make' to build. 00:03:29.725 16:41:51 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:03:29.725 16:41:51 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:03:29.725 16:41:51 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:03:29.725 16:41:51 -- common/autotest_common.sh@10 -- $ set +x 00:03:29.725 ************************************ 00:03:29.725 START TEST make 00:03:29.725 ************************************ 00:03:29.725 16:41:51 make -- common/autotest_common.sh@1125 -- $ make -j10 00:03:30.295 make[1]: Nothing to be done for 'all'. 00:04:16.991 CC lib/ut/ut.o 00:04:16.991 CC lib/ut_mock/mock.o 00:04:16.991 CC lib/log/log_deprecated.o 00:04:16.991 CC lib/log/log.o 00:04:16.991 CC lib/log/log_flags.o 00:04:16.991 LIB libspdk_ut_mock.a 00:04:16.991 LIB libspdk_ut.a 00:04:16.991 LIB libspdk_log.a 00:04:16.991 SO libspdk_ut_mock.so.6.0 00:04:16.991 SO libspdk_ut.so.2.0 00:04:16.991 SYMLINK libspdk_ut_mock.so 00:04:16.991 SO libspdk_log.so.7.0 00:04:16.991 SYMLINK libspdk_ut.so 00:04:16.991 SYMLINK libspdk_log.so 00:04:16.991 CXX lib/trace_parser/trace.o 00:04:16.991 CC lib/util/base64.o 00:04:16.991 CC lib/util/bit_array.o 00:04:16.991 CC lib/util/crc16.o 00:04:16.991 CC lib/util/cpuset.o 00:04:16.991 CC lib/util/crc32.o 00:04:16.991 CC lib/util/crc32c.o 00:04:16.991 CC lib/ioat/ioat.o 00:04:16.991 CC lib/dma/dma.o 00:04:16.991 CC lib/vfio_user/host/vfio_user_pci.o 00:04:16.991 CC lib/util/crc32_ieee.o 00:04:16.991 CC lib/util/crc64.o 00:04:16.991 CC lib/util/dif.o 00:04:16.991 CC lib/util/fd.o 00:04:16.991 CC lib/vfio_user/host/vfio_user.o 00:04:16.991 CC lib/util/fd_group.o 00:04:16.991 LIB libspdk_dma.a 00:04:16.991 SO libspdk_dma.so.5.0 00:04:16.991 CC lib/util/file.o 00:04:16.991 CC lib/util/hexlify.o 00:04:16.991 SYMLINK libspdk_dma.so 00:04:16.991 CC lib/util/iov.o 00:04:16.991 CC lib/util/math.o 00:04:16.991 LIB libspdk_ioat.a 00:04:16.991 CC lib/util/net.o 00:04:16.991 SO libspdk_ioat.so.7.0 00:04:16.991 CC lib/util/pipe.o 00:04:16.991 CC lib/util/strerror_tls.o 00:04:16.991 LIB libspdk_vfio_user.a 00:04:16.991 SYMLINK libspdk_ioat.so 00:04:16.991 CC lib/util/string.o 00:04:16.991 CC lib/util/uuid.o 00:04:16.991 SO libspdk_vfio_user.so.5.0 00:04:16.991 CC lib/util/xor.o 00:04:16.991 SYMLINK libspdk_vfio_user.so 00:04:16.991 CC lib/util/zipf.o 00:04:16.991 CC lib/util/md5.o 00:04:16.991 LIB libspdk_util.a 00:04:16.991 SO libspdk_util.so.10.0 00:04:16.991 LIB libspdk_trace_parser.a 00:04:16.991 SO libspdk_trace_parser.so.6.0 00:04:16.991 SYMLINK libspdk_util.so 00:04:16.991 SYMLINK libspdk_trace_parser.so 00:04:16.991 CC lib/conf/conf.o 00:04:16.991 CC lib/rdma_utils/rdma_utils.o 00:04:16.991 CC lib/idxd/idxd.o 00:04:16.991 CC lib/rdma_provider/common.o 00:04:16.991 CC lib/idxd/idxd_user.o 00:04:16.991 CC lib/rdma_provider/rdma_provider_verbs.o 00:04:16.991 CC lib/idxd/idxd_kernel.o 00:04:16.991 CC lib/json/json_parse.o 00:04:16.991 CC lib/vmd/vmd.o 00:04:16.991 CC lib/env_dpdk/env.o 00:04:16.991 CC lib/json/json_util.o 00:04:16.991 CC lib/json/json_write.o 00:04:16.991 LIB libspdk_rdma_provider.a 00:04:16.991 SO libspdk_rdma_provider.so.6.0 00:04:16.991 LIB libspdk_conf.a 00:04:16.991 CC lib/vmd/led.o 00:04:16.991 SO libspdk_conf.so.6.0 00:04:16.991 CC lib/env_dpdk/memory.o 00:04:16.991 LIB libspdk_rdma_utils.a 00:04:16.991 SYMLINK libspdk_rdma_provider.so 00:04:16.991 CC lib/env_dpdk/pci.o 00:04:16.991 SYMLINK libspdk_conf.so 00:04:16.991 SO libspdk_rdma_utils.so.1.0 00:04:16.991 CC lib/env_dpdk/init.o 00:04:16.991 SYMLINK libspdk_rdma_utils.so 00:04:16.991 CC lib/env_dpdk/threads.o 00:04:16.991 CC lib/env_dpdk/pci_ioat.o 00:04:16.991 CC lib/env_dpdk/pci_virtio.o 00:04:16.991 LIB libspdk_json.a 00:04:16.991 CC lib/env_dpdk/pci_vmd.o 00:04:16.991 CC lib/env_dpdk/pci_idxd.o 00:04:16.991 SO libspdk_json.so.6.0 00:04:16.991 CC lib/env_dpdk/pci_event.o 00:04:16.991 SYMLINK libspdk_json.so 00:04:16.991 CC lib/env_dpdk/sigbus_handler.o 00:04:16.991 CC lib/env_dpdk/pci_dpdk.o 00:04:16.991 CC lib/env_dpdk/pci_dpdk_2207.o 00:04:16.991 CC lib/env_dpdk/pci_dpdk_2211.o 00:04:16.991 LIB libspdk_idxd.a 00:04:16.991 SO libspdk_idxd.so.12.1 00:04:16.991 LIB libspdk_vmd.a 00:04:16.991 SO libspdk_vmd.so.6.0 00:04:16.991 SYMLINK libspdk_idxd.so 00:04:16.991 SYMLINK libspdk_vmd.so 00:04:16.991 CC lib/jsonrpc/jsonrpc_server.o 00:04:16.991 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:04:16.991 CC lib/jsonrpc/jsonrpc_client.o 00:04:16.991 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:04:16.991 LIB libspdk_jsonrpc.a 00:04:16.991 SO libspdk_jsonrpc.so.6.0 00:04:16.991 SYMLINK libspdk_jsonrpc.so 00:04:16.991 LIB libspdk_env_dpdk.a 00:04:16.991 CC lib/rpc/rpc.o 00:04:16.991 SO libspdk_env_dpdk.so.15.0 00:04:16.991 SYMLINK libspdk_env_dpdk.so 00:04:16.991 LIB libspdk_rpc.a 00:04:16.991 SO libspdk_rpc.so.6.0 00:04:16.991 SYMLINK libspdk_rpc.so 00:04:16.991 CC lib/notify/notify.o 00:04:16.991 CC lib/notify/notify_rpc.o 00:04:16.991 CC lib/keyring/keyring.o 00:04:16.991 CC lib/keyring/keyring_rpc.o 00:04:16.991 CC lib/trace/trace.o 00:04:16.991 CC lib/trace/trace_flags.o 00:04:16.991 CC lib/trace/trace_rpc.o 00:04:16.991 LIB libspdk_notify.a 00:04:16.991 SO libspdk_notify.so.6.0 00:04:16.991 LIB libspdk_keyring.a 00:04:16.991 SYMLINK libspdk_notify.so 00:04:16.991 LIB libspdk_trace.a 00:04:16.991 SO libspdk_keyring.so.2.0 00:04:16.991 SO libspdk_trace.so.11.0 00:04:16.991 SYMLINK libspdk_keyring.so 00:04:16.991 SYMLINK libspdk_trace.so 00:04:16.991 CC lib/sock/sock.o 00:04:16.991 CC lib/sock/sock_rpc.o 00:04:16.991 CC lib/thread/iobuf.o 00:04:16.991 CC lib/thread/thread.o 00:04:16.991 LIB libspdk_sock.a 00:04:16.991 SO libspdk_sock.so.10.0 00:04:16.991 SYMLINK libspdk_sock.so 00:04:16.991 CC lib/nvme/nvme_ctrlr_cmd.o 00:04:16.991 CC lib/nvme/nvme_ctrlr.o 00:04:16.991 CC lib/nvme/nvme_fabric.o 00:04:16.991 CC lib/nvme/nvme_ns_cmd.o 00:04:16.991 CC lib/nvme/nvme_ns.o 00:04:16.992 CC lib/nvme/nvme_pcie_common.o 00:04:16.992 CC lib/nvme/nvme_pcie.o 00:04:16.992 CC lib/nvme/nvme_qpair.o 00:04:16.992 CC lib/nvme/nvme.o 00:04:16.992 LIB libspdk_thread.a 00:04:17.251 SO libspdk_thread.so.10.1 00:04:17.251 CC lib/nvme/nvme_quirks.o 00:04:17.251 CC lib/nvme/nvme_transport.o 00:04:17.251 SYMLINK libspdk_thread.so 00:04:17.251 CC lib/nvme/nvme_discovery.o 00:04:17.251 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:04:17.251 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:04:17.510 CC lib/nvme/nvme_tcp.o 00:04:17.510 CC lib/nvme/nvme_opal.o 00:04:17.510 CC lib/nvme/nvme_io_msg.o 00:04:17.510 CC lib/nvme/nvme_poll_group.o 00:04:17.769 CC lib/nvme/nvme_zns.o 00:04:17.769 CC lib/nvme/nvme_stubs.o 00:04:17.769 CC lib/nvme/nvme_auth.o 00:04:17.769 CC lib/nvme/nvme_cuse.o 00:04:17.769 CC lib/nvme/nvme_rdma.o 00:04:18.335 CC lib/accel/accel.o 00:04:18.335 CC lib/accel/accel_rpc.o 00:04:18.335 CC lib/blob/blobstore.o 00:04:18.335 CC lib/init/json_config.o 00:04:18.335 CC lib/virtio/virtio.o 00:04:18.335 CC lib/init/subsystem.o 00:04:18.593 CC lib/init/subsystem_rpc.o 00:04:18.593 CC lib/init/rpc.o 00:04:18.593 CC lib/virtio/virtio_vhost_user.o 00:04:18.593 CC lib/virtio/virtio_vfio_user.o 00:04:18.593 CC lib/virtio/virtio_pci.o 00:04:18.851 LIB libspdk_init.a 00:04:18.851 SO libspdk_init.so.6.0 00:04:18.851 CC lib/blob/request.o 00:04:18.851 SYMLINK libspdk_init.so 00:04:18.851 CC lib/fsdev/fsdev.o 00:04:18.851 CC lib/fsdev/fsdev_io.o 00:04:18.851 CC lib/fsdev/fsdev_rpc.o 00:04:18.851 LIB libspdk_virtio.a 00:04:18.851 CC lib/accel/accel_sw.o 00:04:18.851 CC lib/event/app.o 00:04:19.109 SO libspdk_virtio.so.7.0 00:04:19.109 SYMLINK libspdk_virtio.so 00:04:19.109 CC lib/event/reactor.o 00:04:19.109 CC lib/event/log_rpc.o 00:04:19.109 CC lib/blob/zeroes.o 00:04:19.109 CC lib/event/app_rpc.o 00:04:19.109 CC lib/event/scheduler_static.o 00:04:19.368 LIB libspdk_nvme.a 00:04:19.368 CC lib/blob/blob_bs_dev.o 00:04:19.368 LIB libspdk_accel.a 00:04:19.368 SO libspdk_accel.so.16.0 00:04:19.368 SO libspdk_nvme.so.14.0 00:04:19.368 SYMLINK libspdk_accel.so 00:04:19.368 LIB libspdk_event.a 00:04:19.368 LIB libspdk_fsdev.a 00:04:19.627 SO libspdk_event.so.14.0 00:04:19.627 SO libspdk_fsdev.so.1.0 00:04:19.627 SYMLINK libspdk_event.so 00:04:19.627 SYMLINK libspdk_fsdev.so 00:04:19.627 SYMLINK libspdk_nvme.so 00:04:19.886 CC lib/bdev/bdev.o 00:04:19.886 CC lib/bdev/bdev_rpc.o 00:04:19.886 CC lib/bdev/bdev_zone.o 00:04:19.886 CC lib/bdev/part.o 00:04:19.886 CC lib/bdev/scsi_nvme.o 00:04:19.886 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:04:20.823 LIB libspdk_fuse_dispatcher.a 00:04:20.823 SO libspdk_fuse_dispatcher.so.1.0 00:04:20.823 SYMLINK libspdk_fuse_dispatcher.so 00:04:21.391 LIB libspdk_blob.a 00:04:21.650 SO libspdk_blob.so.11.0 00:04:21.650 SYMLINK libspdk_blob.so 00:04:22.217 CC lib/lvol/lvol.o 00:04:22.217 CC lib/blobfs/blobfs.o 00:04:22.217 CC lib/blobfs/tree.o 00:04:22.476 LIB libspdk_bdev.a 00:04:22.476 SO libspdk_bdev.so.16.0 00:04:22.476 SYMLINK libspdk_bdev.so 00:04:22.736 CC lib/nbd/nbd.o 00:04:22.736 CC lib/nbd/nbd_rpc.o 00:04:22.736 CC lib/ftl/ftl_core.o 00:04:22.736 CC lib/ftl/ftl_init.o 00:04:22.736 CC lib/ftl/ftl_layout.o 00:04:22.736 CC lib/ublk/ublk.o 00:04:22.736 CC lib/nvmf/ctrlr.o 00:04:22.736 CC lib/scsi/dev.o 00:04:22.736 LIB libspdk_blobfs.a 00:04:22.994 SO libspdk_blobfs.so.10.0 00:04:22.994 SYMLINK libspdk_blobfs.so 00:04:22.994 CC lib/scsi/lun.o 00:04:22.994 CC lib/scsi/port.o 00:04:22.994 LIB libspdk_lvol.a 00:04:22.994 CC lib/scsi/scsi.o 00:04:22.994 SO libspdk_lvol.so.10.0 00:04:22.994 SYMLINK libspdk_lvol.so 00:04:22.994 CC lib/ftl/ftl_debug.o 00:04:22.994 CC lib/ftl/ftl_io.o 00:04:22.994 CC lib/ftl/ftl_sb.o 00:04:22.994 CC lib/ftl/ftl_l2p.o 00:04:22.994 CC lib/ftl/ftl_l2p_flat.o 00:04:23.252 CC lib/ublk/ublk_rpc.o 00:04:23.252 LIB libspdk_nbd.a 00:04:23.252 CC lib/scsi/scsi_bdev.o 00:04:23.252 SO libspdk_nbd.so.7.0 00:04:23.252 CC lib/ftl/ftl_nv_cache.o 00:04:23.252 CC lib/ftl/ftl_band.o 00:04:23.252 SYMLINK libspdk_nbd.so 00:04:23.252 CC lib/ftl/ftl_band_ops.o 00:04:23.252 CC lib/ftl/ftl_writer.o 00:04:23.252 CC lib/ftl/ftl_rq.o 00:04:23.252 CC lib/ftl/ftl_reloc.o 00:04:23.252 CC lib/ftl/ftl_l2p_cache.o 00:04:23.510 LIB libspdk_ublk.a 00:04:23.510 CC lib/ftl/ftl_p2l.o 00:04:23.510 SO libspdk_ublk.so.3.0 00:04:23.510 SYMLINK libspdk_ublk.so 00:04:23.510 CC lib/ftl/ftl_p2l_log.o 00:04:23.511 CC lib/ftl/mngt/ftl_mngt.o 00:04:23.511 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:04:23.769 CC lib/scsi/scsi_pr.o 00:04:23.769 CC lib/nvmf/ctrlr_discovery.o 00:04:23.769 CC lib/scsi/scsi_rpc.o 00:04:23.769 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:04:23.769 CC lib/ftl/mngt/ftl_mngt_startup.o 00:04:23.769 CC lib/ftl/mngt/ftl_mngt_md.o 00:04:23.769 CC lib/ftl/mngt/ftl_mngt_misc.o 00:04:23.769 CC lib/scsi/task.o 00:04:23.769 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:04:24.028 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:04:24.028 CC lib/nvmf/ctrlr_bdev.o 00:04:24.028 CC lib/ftl/mngt/ftl_mngt_band.o 00:04:24.028 LIB libspdk_scsi.a 00:04:24.028 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:04:24.028 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:04:24.028 SO libspdk_scsi.so.9.0 00:04:24.028 CC lib/nvmf/subsystem.o 00:04:24.028 CC lib/nvmf/nvmf.o 00:04:24.028 CC lib/nvmf/nvmf_rpc.o 00:04:24.286 SYMLINK libspdk_scsi.so 00:04:24.286 CC lib/nvmf/transport.o 00:04:24.286 CC lib/nvmf/tcp.o 00:04:24.286 CC lib/nvmf/stubs.o 00:04:24.286 CC lib/nvmf/mdns_server.o 00:04:24.286 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:04:24.852 CC lib/nvmf/rdma.o 00:04:24.852 CC lib/nvmf/auth.o 00:04:24.852 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:04:24.852 CC lib/ftl/utils/ftl_conf.o 00:04:24.852 CC lib/iscsi/conn.o 00:04:24.852 CC lib/iscsi/init_grp.o 00:04:24.852 CC lib/iscsi/iscsi.o 00:04:24.852 CC lib/ftl/utils/ftl_md.o 00:04:25.111 CC lib/ftl/utils/ftl_mempool.o 00:04:25.111 CC lib/ftl/utils/ftl_bitmap.o 00:04:25.111 CC lib/ftl/utils/ftl_property.o 00:04:25.111 CC lib/iscsi/param.o 00:04:25.369 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:04:25.369 CC lib/iscsi/portal_grp.o 00:04:25.369 CC lib/iscsi/tgt_node.o 00:04:25.369 CC lib/iscsi/iscsi_subsystem.o 00:04:25.369 CC lib/iscsi/iscsi_rpc.o 00:04:25.369 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:04:25.628 CC lib/iscsi/task.o 00:04:25.628 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:04:25.628 CC lib/vhost/vhost.o 00:04:25.628 CC lib/vhost/vhost_rpc.o 00:04:25.628 CC lib/vhost/vhost_scsi.o 00:04:25.887 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:04:25.887 CC lib/vhost/vhost_blk.o 00:04:25.887 CC lib/vhost/rte_vhost_user.o 00:04:25.887 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:04:25.887 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:04:25.887 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:04:26.145 CC lib/ftl/upgrade/ftl_sb_v3.o 00:04:26.145 CC lib/ftl/upgrade/ftl_sb_v5.o 00:04:26.145 CC lib/ftl/nvc/ftl_nvc_dev.o 00:04:26.145 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:04:26.145 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:04:26.145 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:04:26.404 CC lib/ftl/base/ftl_base_dev.o 00:04:26.404 LIB libspdk_iscsi.a 00:04:26.404 CC lib/ftl/base/ftl_base_bdev.o 00:04:26.404 CC lib/ftl/ftl_trace.o 00:04:26.404 SO libspdk_iscsi.so.8.0 00:04:26.663 SYMLINK libspdk_iscsi.so 00:04:26.663 LIB libspdk_ftl.a 00:04:26.922 LIB libspdk_vhost.a 00:04:26.922 SO libspdk_vhost.so.8.0 00:04:26.922 SO libspdk_ftl.so.9.0 00:04:26.922 LIB libspdk_nvmf.a 00:04:26.922 SYMLINK libspdk_vhost.so 00:04:27.182 SO libspdk_nvmf.so.19.0 00:04:27.182 SYMLINK libspdk_ftl.so 00:04:27.442 SYMLINK libspdk_nvmf.so 00:04:27.700 CC module/env_dpdk/env_dpdk_rpc.o 00:04:27.700 CC module/keyring/linux/keyring.o 00:04:27.700 CC module/scheduler/gscheduler/gscheduler.o 00:04:27.700 CC module/keyring/file/keyring.o 00:04:27.700 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:04:27.700 CC module/scheduler/dynamic/scheduler_dynamic.o 00:04:27.700 CC module/accel/error/accel_error.o 00:04:27.700 CC module/fsdev/aio/fsdev_aio.o 00:04:27.700 CC module/blob/bdev/blob_bdev.o 00:04:27.700 CC module/sock/posix/posix.o 00:04:27.959 LIB libspdk_env_dpdk_rpc.a 00:04:27.959 SO libspdk_env_dpdk_rpc.so.6.0 00:04:27.959 SYMLINK libspdk_env_dpdk_rpc.so 00:04:27.959 CC module/keyring/file/keyring_rpc.o 00:04:27.959 CC module/fsdev/aio/fsdev_aio_rpc.o 00:04:27.959 LIB libspdk_scheduler_gscheduler.a 00:04:27.959 CC module/keyring/linux/keyring_rpc.o 00:04:27.959 LIB libspdk_scheduler_dpdk_governor.a 00:04:27.959 SO libspdk_scheduler_gscheduler.so.4.0 00:04:27.959 SO libspdk_scheduler_dpdk_governor.so.4.0 00:04:27.959 CC module/accel/error/accel_error_rpc.o 00:04:27.959 LIB libspdk_scheduler_dynamic.a 00:04:27.959 SYMLINK libspdk_scheduler_gscheduler.so 00:04:27.959 CC module/fsdev/aio/linux_aio_mgr.o 00:04:27.959 SYMLINK libspdk_scheduler_dpdk_governor.so 00:04:27.959 SO libspdk_scheduler_dynamic.so.4.0 00:04:27.959 LIB libspdk_keyring_file.a 00:04:27.959 LIB libspdk_blob_bdev.a 00:04:27.959 SYMLINK libspdk_scheduler_dynamic.so 00:04:27.959 LIB libspdk_keyring_linux.a 00:04:27.959 SO libspdk_keyring_file.so.2.0 00:04:28.218 SO libspdk_blob_bdev.so.11.0 00:04:28.218 SO libspdk_keyring_linux.so.1.0 00:04:28.218 LIB libspdk_accel_error.a 00:04:28.218 SO libspdk_accel_error.so.2.0 00:04:28.218 SYMLINK libspdk_keyring_file.so 00:04:28.218 SYMLINK libspdk_keyring_linux.so 00:04:28.218 SYMLINK libspdk_blob_bdev.so 00:04:28.218 CC module/accel/ioat/accel_ioat.o 00:04:28.218 SYMLINK libspdk_accel_error.so 00:04:28.218 CC module/accel/ioat/accel_ioat_rpc.o 00:04:28.218 CC module/accel/dsa/accel_dsa.o 00:04:28.218 CC module/accel/dsa/accel_dsa_rpc.o 00:04:28.218 CC module/accel/iaa/accel_iaa.o 00:04:28.218 CC module/accel/iaa/accel_iaa_rpc.o 00:04:28.478 LIB libspdk_accel_ioat.a 00:04:28.478 CC module/blobfs/bdev/blobfs_bdev.o 00:04:28.478 SO libspdk_accel_ioat.so.6.0 00:04:28.478 CC module/bdev/delay/vbdev_delay.o 00:04:28.478 LIB libspdk_accel_iaa.a 00:04:28.478 SYMLINK libspdk_accel_ioat.so 00:04:28.478 CC module/bdev/delay/vbdev_delay_rpc.o 00:04:28.478 SO libspdk_accel_iaa.so.3.0 00:04:28.478 CC module/bdev/error/vbdev_error.o 00:04:28.478 LIB libspdk_fsdev_aio.a 00:04:28.478 CC module/bdev/lvol/vbdev_lvol.o 00:04:28.478 CC module/bdev/gpt/gpt.o 00:04:28.478 LIB libspdk_accel_dsa.a 00:04:28.478 SYMLINK libspdk_accel_iaa.so 00:04:28.478 SO libspdk_fsdev_aio.so.1.0 00:04:28.478 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:04:28.478 SO libspdk_accel_dsa.so.5.0 00:04:28.478 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:04:28.478 LIB libspdk_sock_posix.a 00:04:28.478 SYMLINK libspdk_fsdev_aio.so 00:04:28.478 CC module/bdev/error/vbdev_error_rpc.o 00:04:28.478 SYMLINK libspdk_accel_dsa.so 00:04:28.478 CC module/bdev/gpt/vbdev_gpt.o 00:04:28.478 SO libspdk_sock_posix.so.6.0 00:04:28.737 SYMLINK libspdk_sock_posix.so 00:04:28.737 LIB libspdk_blobfs_bdev.a 00:04:28.737 SO libspdk_blobfs_bdev.so.6.0 00:04:28.737 LIB libspdk_bdev_error.a 00:04:28.737 CC module/bdev/malloc/bdev_malloc.o 00:04:28.737 LIB libspdk_bdev_delay.a 00:04:28.737 SO libspdk_bdev_error.so.6.0 00:04:28.737 SO libspdk_bdev_delay.so.6.0 00:04:28.737 SYMLINK libspdk_blobfs_bdev.so 00:04:28.737 CC module/bdev/null/bdev_null.o 00:04:28.737 CC module/bdev/null/bdev_null_rpc.o 00:04:28.737 CC module/bdev/nvme/bdev_nvme.o 00:04:28.737 SYMLINK libspdk_bdev_error.so 00:04:28.737 SYMLINK libspdk_bdev_delay.so 00:04:28.737 CC module/bdev/passthru/vbdev_passthru.o 00:04:28.737 LIB libspdk_bdev_gpt.a 00:04:28.996 SO libspdk_bdev_gpt.so.6.0 00:04:28.996 SYMLINK libspdk_bdev_gpt.so 00:04:28.996 CC module/bdev/malloc/bdev_malloc_rpc.o 00:04:28.996 CC module/bdev/split/vbdev_split.o 00:04:28.996 CC module/bdev/raid/bdev_raid.o 00:04:28.996 LIB libspdk_bdev_lvol.a 00:04:28.996 CC module/bdev/zone_block/vbdev_zone_block.o 00:04:28.996 SO libspdk_bdev_lvol.so.6.0 00:04:28.996 LIB libspdk_bdev_null.a 00:04:28.996 CC module/bdev/aio/bdev_aio.o 00:04:28.996 SO libspdk_bdev_null.so.6.0 00:04:28.996 CC module/bdev/split/vbdev_split_rpc.o 00:04:28.996 SYMLINK libspdk_bdev_lvol.so 00:04:28.996 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:04:29.255 LIB libspdk_bdev_malloc.a 00:04:29.255 SYMLINK libspdk_bdev_null.so 00:04:29.255 CC module/bdev/raid/bdev_raid_rpc.o 00:04:29.255 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:04:29.255 SO libspdk_bdev_malloc.so.6.0 00:04:29.255 CC module/bdev/raid/bdev_raid_sb.o 00:04:29.255 SYMLINK libspdk_bdev_malloc.so 00:04:29.255 CC module/bdev/raid/raid0.o 00:04:29.255 LIB libspdk_bdev_split.a 00:04:29.255 SO libspdk_bdev_split.so.6.0 00:04:29.255 LIB libspdk_bdev_passthru.a 00:04:29.255 SO libspdk_bdev_passthru.so.6.0 00:04:29.255 SYMLINK libspdk_bdev_split.so 00:04:29.255 CC module/bdev/aio/bdev_aio_rpc.o 00:04:29.255 CC module/bdev/nvme/bdev_nvme_rpc.o 00:04:29.255 LIB libspdk_bdev_zone_block.a 00:04:29.255 SYMLINK libspdk_bdev_passthru.so 00:04:29.514 SO libspdk_bdev_zone_block.so.6.0 00:04:29.514 CC module/bdev/ftl/bdev_ftl.o 00:04:29.514 CC module/bdev/raid/raid1.o 00:04:29.514 CC module/bdev/raid/concat.o 00:04:29.514 SYMLINK libspdk_bdev_zone_block.so 00:04:29.514 CC module/bdev/raid/raid5f.o 00:04:29.514 LIB libspdk_bdev_aio.a 00:04:29.514 SO libspdk_bdev_aio.so.6.0 00:04:29.514 CC module/bdev/iscsi/bdev_iscsi.o 00:04:29.773 SYMLINK libspdk_bdev_aio.so 00:04:29.773 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:04:29.773 CC module/bdev/virtio/bdev_virtio_scsi.o 00:04:29.773 CC module/bdev/virtio/bdev_virtio_blk.o 00:04:29.773 CC module/bdev/ftl/bdev_ftl_rpc.o 00:04:29.773 CC module/bdev/nvme/nvme_rpc.o 00:04:29.773 CC module/bdev/nvme/bdev_mdns_client.o 00:04:29.773 LIB libspdk_bdev_ftl.a 00:04:30.032 SO libspdk_bdev_ftl.so.6.0 00:04:30.032 CC module/bdev/nvme/vbdev_opal.o 00:04:30.032 LIB libspdk_bdev_iscsi.a 00:04:30.032 CC module/bdev/virtio/bdev_virtio_rpc.o 00:04:30.032 SYMLINK libspdk_bdev_ftl.so 00:04:30.032 SO libspdk_bdev_iscsi.so.6.0 00:04:30.032 CC module/bdev/nvme/vbdev_opal_rpc.o 00:04:30.032 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:04:30.032 SYMLINK libspdk_bdev_iscsi.so 00:04:30.032 LIB libspdk_bdev_raid.a 00:04:30.032 SO libspdk_bdev_raid.so.6.0 00:04:30.293 LIB libspdk_bdev_virtio.a 00:04:30.293 SO libspdk_bdev_virtio.so.6.0 00:04:30.293 SYMLINK libspdk_bdev_raid.so 00:04:30.293 SYMLINK libspdk_bdev_virtio.so 00:04:31.232 LIB libspdk_bdev_nvme.a 00:04:31.232 SO libspdk_bdev_nvme.so.7.0 00:04:31.232 SYMLINK libspdk_bdev_nvme.so 00:04:31.802 CC module/event/subsystems/keyring/keyring.o 00:04:31.802 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:04:31.802 CC module/event/subsystems/vmd/vmd.o 00:04:31.802 CC module/event/subsystems/vmd/vmd_rpc.o 00:04:31.802 CC module/event/subsystems/scheduler/scheduler.o 00:04:31.802 CC module/event/subsystems/sock/sock.o 00:04:31.802 CC module/event/subsystems/fsdev/fsdev.o 00:04:31.802 CC module/event/subsystems/iobuf/iobuf.o 00:04:31.802 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:04:32.061 LIB libspdk_event_keyring.a 00:04:32.061 LIB libspdk_event_scheduler.a 00:04:32.061 LIB libspdk_event_sock.a 00:04:32.061 LIB libspdk_event_fsdev.a 00:04:32.061 LIB libspdk_event_vmd.a 00:04:32.061 LIB libspdk_event_vhost_blk.a 00:04:32.061 SO libspdk_event_scheduler.so.4.0 00:04:32.061 SO libspdk_event_fsdev.so.1.0 00:04:32.061 SO libspdk_event_keyring.so.1.0 00:04:32.061 SO libspdk_event_sock.so.5.0 00:04:32.061 SO libspdk_event_vmd.so.6.0 00:04:32.061 LIB libspdk_event_iobuf.a 00:04:32.061 SO libspdk_event_vhost_blk.so.3.0 00:04:32.061 SO libspdk_event_iobuf.so.3.0 00:04:32.061 SYMLINK libspdk_event_fsdev.so 00:04:32.061 SYMLINK libspdk_event_scheduler.so 00:04:32.061 SYMLINK libspdk_event_keyring.so 00:04:32.061 SYMLINK libspdk_event_sock.so 00:04:32.061 SYMLINK libspdk_event_vmd.so 00:04:32.061 SYMLINK libspdk_event_vhost_blk.so 00:04:32.061 SYMLINK libspdk_event_iobuf.so 00:04:32.631 CC module/event/subsystems/accel/accel.o 00:04:32.631 LIB libspdk_event_accel.a 00:04:32.631 SO libspdk_event_accel.so.6.0 00:04:32.631 SYMLINK libspdk_event_accel.so 00:04:33.201 CC module/event/subsystems/bdev/bdev.o 00:04:33.201 LIB libspdk_event_bdev.a 00:04:33.201 SO libspdk_event_bdev.so.6.0 00:04:33.460 SYMLINK libspdk_event_bdev.so 00:04:33.720 CC module/event/subsystems/scsi/scsi.o 00:04:33.720 CC module/event/subsystems/ublk/ublk.o 00:04:33.720 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:04:33.720 CC module/event/subsystems/nbd/nbd.o 00:04:33.720 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:04:33.986 LIB libspdk_event_ublk.a 00:04:33.986 LIB libspdk_event_nbd.a 00:04:33.986 SO libspdk_event_ublk.so.3.0 00:04:33.986 LIB libspdk_event_scsi.a 00:04:33.986 SO libspdk_event_nbd.so.6.0 00:04:33.986 SYMLINK libspdk_event_ublk.so 00:04:33.986 SO libspdk_event_scsi.so.6.0 00:04:33.986 SYMLINK libspdk_event_nbd.so 00:04:33.986 LIB libspdk_event_nvmf.a 00:04:33.986 SO libspdk_event_nvmf.so.6.0 00:04:33.986 SYMLINK libspdk_event_scsi.so 00:04:33.986 SYMLINK libspdk_event_nvmf.so 00:04:34.264 CC module/event/subsystems/iscsi/iscsi.o 00:04:34.264 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:04:34.544 LIB libspdk_event_vhost_scsi.a 00:04:34.544 LIB libspdk_event_iscsi.a 00:04:34.544 SO libspdk_event_vhost_scsi.so.3.0 00:04:34.544 SO libspdk_event_iscsi.so.6.0 00:04:34.544 SYMLINK libspdk_event_vhost_scsi.so 00:04:34.544 SYMLINK libspdk_event_iscsi.so 00:04:34.804 SO libspdk.so.6.0 00:04:34.804 SYMLINK libspdk.so 00:04:35.063 CC app/trace_record/trace_record.o 00:04:35.063 CXX app/trace/trace.o 00:04:35.063 CC app/spdk_lspci/spdk_lspci.o 00:04:35.063 CC examples/interrupt_tgt/interrupt_tgt.o 00:04:35.063 CC app/iscsi_tgt/iscsi_tgt.o 00:04:35.063 CC app/nvmf_tgt/nvmf_main.o 00:04:35.063 CC examples/util/zipf/zipf.o 00:04:35.063 CC app/spdk_tgt/spdk_tgt.o 00:04:35.322 CC test/thread/poller_perf/poller_perf.o 00:04:35.322 CC examples/ioat/perf/perf.o 00:04:35.322 LINK spdk_lspci 00:04:35.322 LINK interrupt_tgt 00:04:35.322 LINK zipf 00:04:35.322 LINK poller_perf 00:04:35.322 LINK nvmf_tgt 00:04:35.322 LINK iscsi_tgt 00:04:35.322 LINK spdk_trace_record 00:04:35.322 LINK spdk_tgt 00:04:35.322 LINK ioat_perf 00:04:35.580 CC app/spdk_nvme_perf/perf.o 00:04:35.580 LINK spdk_trace 00:04:35.580 CC app/spdk_nvme_identify/identify.o 00:04:35.580 CC app/spdk_nvme_discover/discovery_aer.o 00:04:35.580 CC app/spdk_top/spdk_top.o 00:04:35.580 CC examples/ioat/verify/verify.o 00:04:35.580 CC app/spdk_dd/spdk_dd.o 00:04:35.580 CC test/dma/test_dma/test_dma.o 00:04:35.580 CC app/fio/nvme/fio_plugin.o 00:04:35.839 CC examples/thread/thread/thread_ex.o 00:04:35.839 LINK spdk_nvme_discover 00:04:35.839 CC app/fio/bdev/fio_plugin.o 00:04:35.839 LINK verify 00:04:36.098 LINK spdk_dd 00:04:36.098 LINK thread 00:04:36.098 CC app/vhost/vhost.o 00:04:36.098 LINK test_dma 00:04:36.098 CC examples/sock/hello_world/hello_sock.o 00:04:36.098 LINK vhost 00:04:36.355 LINK spdk_bdev 00:04:36.355 LINK spdk_nvme 00:04:36.355 CC examples/vmd/lsvmd/lsvmd.o 00:04:36.355 CC examples/idxd/perf/perf.o 00:04:36.355 LINK spdk_nvme_perf 00:04:36.355 LINK hello_sock 00:04:36.355 LINK spdk_nvme_identify 00:04:36.355 LINK lsvmd 00:04:36.355 CC examples/vmd/led/led.o 00:04:36.614 CC test/app/bdev_svc/bdev_svc.o 00:04:36.614 CC examples/accel/perf/accel_perf.o 00:04:36.614 LINK spdk_top 00:04:36.614 CC examples/fsdev/hello_world/hello_fsdev.o 00:04:36.614 LINK led 00:04:36.614 LINK idxd_perf 00:04:36.614 LINK bdev_svc 00:04:36.614 CC examples/blob/hello_world/hello_blob.o 00:04:36.614 CC test/blobfs/mkfs/mkfs.o 00:04:36.614 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:04:36.872 TEST_HEADER include/spdk/accel.h 00:04:36.872 TEST_HEADER include/spdk/accel_module.h 00:04:36.872 TEST_HEADER include/spdk/assert.h 00:04:36.872 CC examples/nvme/hello_world/hello_world.o 00:04:36.872 TEST_HEADER include/spdk/barrier.h 00:04:36.872 TEST_HEADER include/spdk/base64.h 00:04:36.872 TEST_HEADER include/spdk/bdev.h 00:04:36.872 TEST_HEADER include/spdk/bdev_module.h 00:04:36.872 TEST_HEADER include/spdk/bdev_zone.h 00:04:36.872 TEST_HEADER include/spdk/bit_array.h 00:04:36.872 TEST_HEADER include/spdk/bit_pool.h 00:04:36.872 TEST_HEADER include/spdk/blob_bdev.h 00:04:36.873 TEST_HEADER include/spdk/blobfs_bdev.h 00:04:36.873 TEST_HEADER include/spdk/blobfs.h 00:04:36.873 TEST_HEADER include/spdk/blob.h 00:04:36.873 TEST_HEADER include/spdk/conf.h 00:04:36.873 TEST_HEADER include/spdk/config.h 00:04:36.873 TEST_HEADER include/spdk/cpuset.h 00:04:36.873 TEST_HEADER include/spdk/crc16.h 00:04:36.873 TEST_HEADER include/spdk/crc32.h 00:04:36.873 TEST_HEADER include/spdk/crc64.h 00:04:36.873 TEST_HEADER include/spdk/dif.h 00:04:36.873 TEST_HEADER include/spdk/dma.h 00:04:36.873 TEST_HEADER include/spdk/endian.h 00:04:36.873 TEST_HEADER include/spdk/env_dpdk.h 00:04:36.873 TEST_HEADER include/spdk/env.h 00:04:36.873 TEST_HEADER include/spdk/event.h 00:04:36.873 TEST_HEADER include/spdk/fd_group.h 00:04:36.873 TEST_HEADER include/spdk/fd.h 00:04:36.873 LINK hello_fsdev 00:04:36.873 TEST_HEADER include/spdk/file.h 00:04:36.873 TEST_HEADER include/spdk/fsdev.h 00:04:36.873 TEST_HEADER include/spdk/fsdev_module.h 00:04:36.873 TEST_HEADER include/spdk/ftl.h 00:04:36.873 TEST_HEADER include/spdk/fuse_dispatcher.h 00:04:36.873 TEST_HEADER include/spdk/gpt_spec.h 00:04:36.873 TEST_HEADER include/spdk/hexlify.h 00:04:36.873 TEST_HEADER include/spdk/histogram_data.h 00:04:36.873 TEST_HEADER include/spdk/idxd.h 00:04:36.873 TEST_HEADER include/spdk/idxd_spec.h 00:04:36.873 TEST_HEADER include/spdk/init.h 00:04:36.873 TEST_HEADER include/spdk/ioat.h 00:04:36.873 TEST_HEADER include/spdk/ioat_spec.h 00:04:36.873 TEST_HEADER include/spdk/iscsi_spec.h 00:04:36.873 TEST_HEADER include/spdk/json.h 00:04:36.873 TEST_HEADER include/spdk/jsonrpc.h 00:04:36.873 TEST_HEADER include/spdk/keyring.h 00:04:36.873 TEST_HEADER include/spdk/keyring_module.h 00:04:36.873 TEST_HEADER include/spdk/likely.h 00:04:36.873 TEST_HEADER include/spdk/log.h 00:04:36.873 TEST_HEADER include/spdk/lvol.h 00:04:36.873 TEST_HEADER include/spdk/md5.h 00:04:36.873 TEST_HEADER include/spdk/memory.h 00:04:36.873 TEST_HEADER include/spdk/mmio.h 00:04:36.873 TEST_HEADER include/spdk/nbd.h 00:04:36.873 TEST_HEADER include/spdk/net.h 00:04:36.873 TEST_HEADER include/spdk/notify.h 00:04:36.873 TEST_HEADER include/spdk/nvme.h 00:04:36.873 TEST_HEADER include/spdk/nvme_intel.h 00:04:36.873 TEST_HEADER include/spdk/nvme_ocssd.h 00:04:36.873 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:04:36.873 TEST_HEADER include/spdk/nvme_spec.h 00:04:36.873 TEST_HEADER include/spdk/nvme_zns.h 00:04:36.873 TEST_HEADER include/spdk/nvmf_cmd.h 00:04:36.873 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:04:36.873 TEST_HEADER include/spdk/nvmf.h 00:04:36.873 TEST_HEADER include/spdk/nvmf_spec.h 00:04:36.873 TEST_HEADER include/spdk/nvmf_transport.h 00:04:36.873 TEST_HEADER include/spdk/opal.h 00:04:36.873 TEST_HEADER include/spdk/opal_spec.h 00:04:36.873 TEST_HEADER include/spdk/pci_ids.h 00:04:36.873 TEST_HEADER include/spdk/pipe.h 00:04:36.873 TEST_HEADER include/spdk/queue.h 00:04:36.873 TEST_HEADER include/spdk/reduce.h 00:04:36.873 TEST_HEADER include/spdk/rpc.h 00:04:36.873 TEST_HEADER include/spdk/scheduler.h 00:04:36.873 TEST_HEADER include/spdk/scsi.h 00:04:36.873 TEST_HEADER include/spdk/scsi_spec.h 00:04:36.873 TEST_HEADER include/spdk/sock.h 00:04:36.873 TEST_HEADER include/spdk/stdinc.h 00:04:36.873 TEST_HEADER include/spdk/string.h 00:04:36.873 TEST_HEADER include/spdk/thread.h 00:04:36.873 TEST_HEADER include/spdk/trace.h 00:04:36.873 TEST_HEADER include/spdk/trace_parser.h 00:04:36.873 LINK mkfs 00:04:36.873 TEST_HEADER include/spdk/tree.h 00:04:36.873 TEST_HEADER include/spdk/ublk.h 00:04:36.873 TEST_HEADER include/spdk/util.h 00:04:36.873 TEST_HEADER include/spdk/uuid.h 00:04:36.873 TEST_HEADER include/spdk/version.h 00:04:36.873 TEST_HEADER include/spdk/vfio_user_pci.h 00:04:36.873 TEST_HEADER include/spdk/vfio_user_spec.h 00:04:36.873 TEST_HEADER include/spdk/vhost.h 00:04:36.873 TEST_HEADER include/spdk/vmd.h 00:04:36.873 TEST_HEADER include/spdk/xor.h 00:04:36.873 TEST_HEADER include/spdk/zipf.h 00:04:36.873 CXX test/cpp_headers/accel.o 00:04:36.873 CC test/env/vtophys/vtophys.o 00:04:36.873 LINK hello_blob 00:04:36.873 CC test/event/event_perf/event_perf.o 00:04:36.873 CC test/env/mem_callbacks/mem_callbacks.o 00:04:36.873 LINK hello_world 00:04:37.132 LINK accel_perf 00:04:37.132 CC test/event/reactor/reactor.o 00:04:37.132 CXX test/cpp_headers/accel_module.o 00:04:37.132 LINK vtophys 00:04:37.132 LINK event_perf 00:04:37.132 LINK mem_callbacks 00:04:37.132 CC test/event/reactor_perf/reactor_perf.o 00:04:37.132 LINK nvme_fuzz 00:04:37.132 LINK reactor 00:04:37.132 CC examples/nvme/reconnect/reconnect.o 00:04:37.132 CXX test/cpp_headers/assert.o 00:04:37.132 CC examples/blob/cli/blobcli.o 00:04:37.391 CC test/event/app_repeat/app_repeat.o 00:04:37.391 LINK reactor_perf 00:04:37.391 CC test/event/scheduler/scheduler.o 00:04:37.391 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:04:37.391 CXX test/cpp_headers/barrier.o 00:04:37.391 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:04:37.391 CC test/env/memory/memory_ut.o 00:04:37.391 LINK app_repeat 00:04:37.391 CC test/lvol/esnap/esnap.o 00:04:37.391 LINK env_dpdk_post_init 00:04:37.391 CXX test/cpp_headers/base64.o 00:04:37.649 LINK reconnect 00:04:37.649 LINK scheduler 00:04:37.649 CC examples/bdev/hello_world/hello_bdev.o 00:04:37.649 CXX test/cpp_headers/bdev.o 00:04:37.649 CXX test/cpp_headers/bdev_module.o 00:04:37.649 CC examples/bdev/bdevperf/bdevperf.o 00:04:37.649 CXX test/cpp_headers/bdev_zone.o 00:04:37.649 LINK blobcli 00:04:37.649 CC examples/nvme/nvme_manage/nvme_manage.o 00:04:37.907 LINK hello_bdev 00:04:37.907 CC test/env/pci/pci_ut.o 00:04:37.907 CXX test/cpp_headers/bit_array.o 00:04:37.907 CC test/nvme/aer/aer.o 00:04:37.907 CXX test/cpp_headers/bit_pool.o 00:04:37.907 CC test/nvme/reset/reset.o 00:04:38.166 CC test/rpc_client/rpc_client_test.o 00:04:38.166 CXX test/cpp_headers/blob_bdev.o 00:04:38.166 LINK memory_ut 00:04:38.166 LINK aer 00:04:38.166 LINK reset 00:04:38.166 LINK nvme_manage 00:04:38.166 LINK pci_ut 00:04:38.425 CXX test/cpp_headers/blobfs_bdev.o 00:04:38.425 LINK rpc_client_test 00:04:38.425 CXX test/cpp_headers/blobfs.o 00:04:38.425 CXX test/cpp_headers/blob.o 00:04:38.425 CC test/nvme/sgl/sgl.o 00:04:38.425 CC test/accel/dif/dif.o 00:04:38.425 CC examples/nvme/arbitration/arbitration.o 00:04:38.425 LINK bdevperf 00:04:38.684 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:04:38.684 CC test/nvme/e2edp/nvme_dp.o 00:04:38.684 CC test/nvme/overhead/overhead.o 00:04:38.684 CXX test/cpp_headers/conf.o 00:04:38.684 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:04:38.684 LINK sgl 00:04:38.684 CXX test/cpp_headers/config.o 00:04:38.684 CXX test/cpp_headers/cpuset.o 00:04:38.684 CC test/nvme/err_injection/err_injection.o 00:04:38.942 LINK arbitration 00:04:38.942 LINK nvme_dp 00:04:38.942 LINK overhead 00:04:38.942 CXX test/cpp_headers/crc16.o 00:04:38.942 LINK err_injection 00:04:38.942 CC test/nvme/startup/startup.o 00:04:38.942 CC examples/nvme/hotplug/hotplug.o 00:04:38.942 CXX test/cpp_headers/crc32.o 00:04:39.201 LINK vhost_fuzz 00:04:39.201 CC test/nvme/reserve/reserve.o 00:04:39.201 CC test/app/histogram_perf/histogram_perf.o 00:04:39.201 LINK startup 00:04:39.201 LINK iscsi_fuzz 00:04:39.201 CC test/nvme/simple_copy/simple_copy.o 00:04:39.201 LINK dif 00:04:39.201 CXX test/cpp_headers/crc64.o 00:04:39.201 LINK histogram_perf 00:04:39.201 LINK hotplug 00:04:39.201 LINK reserve 00:04:39.201 CC test/nvme/connect_stress/connect_stress.o 00:04:39.461 CXX test/cpp_headers/dif.o 00:04:39.461 CC test/nvme/boot_partition/boot_partition.o 00:04:39.461 CXX test/cpp_headers/dma.o 00:04:39.461 LINK simple_copy 00:04:39.461 CC test/nvme/compliance/nvme_compliance.o 00:04:39.461 CXX test/cpp_headers/endian.o 00:04:39.461 LINK connect_stress 00:04:39.461 CC test/app/jsoncat/jsoncat.o 00:04:39.461 CC examples/nvme/cmb_copy/cmb_copy.o 00:04:39.461 LINK boot_partition 00:04:39.720 CXX test/cpp_headers/env_dpdk.o 00:04:39.720 CC examples/nvme/abort/abort.o 00:04:39.720 LINK jsoncat 00:04:39.720 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:04:39.720 CC test/nvme/fused_ordering/fused_ordering.o 00:04:39.720 LINK cmb_copy 00:04:39.720 CXX test/cpp_headers/env.o 00:04:39.720 CC test/nvme/doorbell_aers/doorbell_aers.o 00:04:39.720 LINK nvme_compliance 00:04:39.720 CC test/nvme/fdp/fdp.o 00:04:39.979 LINK pmr_persistence 00:04:39.979 CC test/app/stub/stub.o 00:04:39.979 CXX test/cpp_headers/event.o 00:04:39.979 LINK fused_ordering 00:04:39.979 CXX test/cpp_headers/fd_group.o 00:04:39.979 LINK doorbell_aers 00:04:39.979 CC test/nvme/cuse/cuse.o 00:04:39.979 CXX test/cpp_headers/fd.o 00:04:39.979 LINK abort 00:04:39.979 CXX test/cpp_headers/file.o 00:04:39.979 CXX test/cpp_headers/fsdev.o 00:04:39.979 LINK stub 00:04:40.237 CXX test/cpp_headers/fsdev_module.o 00:04:40.237 CXX test/cpp_headers/ftl.o 00:04:40.237 LINK fdp 00:04:40.237 CXX test/cpp_headers/fuse_dispatcher.o 00:04:40.237 CXX test/cpp_headers/gpt_spec.o 00:04:40.237 CXX test/cpp_headers/hexlify.o 00:04:40.237 CXX test/cpp_headers/histogram_data.o 00:04:40.237 CC test/bdev/bdevio/bdevio.o 00:04:40.237 CXX test/cpp_headers/idxd.o 00:04:40.237 CXX test/cpp_headers/idxd_spec.o 00:04:40.237 CXX test/cpp_headers/init.o 00:04:40.237 CXX test/cpp_headers/ioat.o 00:04:40.237 CC examples/nvmf/nvmf/nvmf.o 00:04:40.237 CXX test/cpp_headers/ioat_spec.o 00:04:40.494 CXX test/cpp_headers/iscsi_spec.o 00:04:40.494 CXX test/cpp_headers/json.o 00:04:40.494 CXX test/cpp_headers/jsonrpc.o 00:04:40.494 CXX test/cpp_headers/keyring.o 00:04:40.494 CXX test/cpp_headers/keyring_module.o 00:04:40.494 CXX test/cpp_headers/likely.o 00:04:40.494 CXX test/cpp_headers/log.o 00:04:40.494 CXX test/cpp_headers/lvol.o 00:04:40.494 CXX test/cpp_headers/md5.o 00:04:40.494 CXX test/cpp_headers/memory.o 00:04:40.494 CXX test/cpp_headers/mmio.o 00:04:40.494 LINK nvmf 00:04:40.751 CXX test/cpp_headers/nbd.o 00:04:40.751 LINK bdevio 00:04:40.751 CXX test/cpp_headers/net.o 00:04:40.751 CXX test/cpp_headers/notify.o 00:04:40.751 CXX test/cpp_headers/nvme.o 00:04:40.751 CXX test/cpp_headers/nvme_intel.o 00:04:40.751 CXX test/cpp_headers/nvme_ocssd.o 00:04:40.751 CXX test/cpp_headers/nvme_ocssd_spec.o 00:04:40.751 CXX test/cpp_headers/nvme_spec.o 00:04:40.751 CXX test/cpp_headers/nvme_zns.o 00:04:40.751 CXX test/cpp_headers/nvmf_cmd.o 00:04:40.751 CXX test/cpp_headers/nvmf_fc_spec.o 00:04:41.013 CXX test/cpp_headers/nvmf.o 00:04:41.013 CXX test/cpp_headers/nvmf_spec.o 00:04:41.013 CXX test/cpp_headers/nvmf_transport.o 00:04:41.013 CXX test/cpp_headers/opal.o 00:04:41.013 CXX test/cpp_headers/opal_spec.o 00:04:41.013 CXX test/cpp_headers/pci_ids.o 00:04:41.013 CXX test/cpp_headers/pipe.o 00:04:41.013 CXX test/cpp_headers/queue.o 00:04:41.013 CXX test/cpp_headers/reduce.o 00:04:41.013 CXX test/cpp_headers/rpc.o 00:04:41.013 CXX test/cpp_headers/scheduler.o 00:04:41.013 CXX test/cpp_headers/scsi.o 00:04:41.013 CXX test/cpp_headers/scsi_spec.o 00:04:41.013 CXX test/cpp_headers/sock.o 00:04:41.013 CXX test/cpp_headers/stdinc.o 00:04:41.013 CXX test/cpp_headers/string.o 00:04:41.272 CXX test/cpp_headers/thread.o 00:04:41.272 CXX test/cpp_headers/trace.o 00:04:41.272 LINK cuse 00:04:41.272 CXX test/cpp_headers/trace_parser.o 00:04:41.272 CXX test/cpp_headers/tree.o 00:04:41.272 CXX test/cpp_headers/ublk.o 00:04:41.272 CXX test/cpp_headers/util.o 00:04:41.272 CXX test/cpp_headers/uuid.o 00:04:41.272 CXX test/cpp_headers/version.o 00:04:41.272 CXX test/cpp_headers/vfio_user_pci.o 00:04:41.272 CXX test/cpp_headers/vfio_user_spec.o 00:04:41.272 CXX test/cpp_headers/vhost.o 00:04:41.272 CXX test/cpp_headers/vmd.o 00:04:41.272 CXX test/cpp_headers/xor.o 00:04:41.272 CXX test/cpp_headers/zipf.o 00:04:43.176 LINK esnap 00:04:43.176 00:04:43.176 real 1m13.443s 00:04:43.176 user 5m36.730s 00:04:43.176 sys 1m4.018s 00:04:43.176 16:43:04 make -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:04:43.176 16:43:04 make -- common/autotest_common.sh@10 -- $ set +x 00:04:43.176 ************************************ 00:04:43.176 END TEST make 00:04:43.176 ************************************ 00:04:43.176 16:43:04 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:04:43.176 16:43:04 -- pm/common@29 -- $ signal_monitor_resources TERM 00:04:43.176 16:43:04 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:04:43.176 16:43:04 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:43.176 16:43:04 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:04:43.176 16:43:04 -- pm/common@44 -- $ pid=6188 00:04:43.176 16:43:04 -- pm/common@50 -- $ kill -TERM 6188 00:04:43.176 16:43:04 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:43.176 16:43:04 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:04:43.176 16:43:04 -- pm/common@44 -- $ pid=6190 00:04:43.176 16:43:04 -- pm/common@50 -- $ kill -TERM 6190 00:04:43.435 16:43:04 -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:04:43.435 16:43:04 -- common/autotest_common.sh@1681 -- # lcov --version 00:04:43.436 16:43:04 -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:04:43.436 16:43:04 -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:04:43.436 16:43:04 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:04:43.436 16:43:04 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:04:43.436 16:43:04 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:04:43.436 16:43:04 -- scripts/common.sh@336 -- # IFS=.-: 00:04:43.436 16:43:04 -- scripts/common.sh@336 -- # read -ra ver1 00:04:43.436 16:43:04 -- scripts/common.sh@337 -- # IFS=.-: 00:04:43.436 16:43:04 -- scripts/common.sh@337 -- # read -ra ver2 00:04:43.436 16:43:04 -- scripts/common.sh@338 -- # local 'op=<' 00:04:43.436 16:43:04 -- scripts/common.sh@340 -- # ver1_l=2 00:04:43.436 16:43:04 -- scripts/common.sh@341 -- # ver2_l=1 00:04:43.436 16:43:04 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:04:43.436 16:43:04 -- scripts/common.sh@344 -- # case "$op" in 00:04:43.436 16:43:04 -- scripts/common.sh@345 -- # : 1 00:04:43.436 16:43:04 -- scripts/common.sh@364 -- # (( v = 0 )) 00:04:43.436 16:43:04 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:43.436 16:43:04 -- scripts/common.sh@365 -- # decimal 1 00:04:43.436 16:43:04 -- scripts/common.sh@353 -- # local d=1 00:04:43.436 16:43:04 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:04:43.436 16:43:04 -- scripts/common.sh@355 -- # echo 1 00:04:43.436 16:43:04 -- scripts/common.sh@365 -- # ver1[v]=1 00:04:43.436 16:43:04 -- scripts/common.sh@366 -- # decimal 2 00:04:43.436 16:43:04 -- scripts/common.sh@353 -- # local d=2 00:04:43.436 16:43:04 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:04:43.436 16:43:04 -- scripts/common.sh@355 -- # echo 2 00:04:43.436 16:43:04 -- scripts/common.sh@366 -- # ver2[v]=2 00:04:43.436 16:43:04 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:04:43.436 16:43:04 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:04:43.436 16:43:04 -- scripts/common.sh@368 -- # return 0 00:04:43.436 16:43:04 -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:04:43.436 16:43:04 -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:04:43.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:43.436 --rc genhtml_branch_coverage=1 00:04:43.436 --rc genhtml_function_coverage=1 00:04:43.436 --rc genhtml_legend=1 00:04:43.436 --rc geninfo_all_blocks=1 00:04:43.436 --rc geninfo_unexecuted_blocks=1 00:04:43.436 00:04:43.436 ' 00:04:43.436 16:43:04 -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:04:43.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:43.436 --rc genhtml_branch_coverage=1 00:04:43.436 --rc genhtml_function_coverage=1 00:04:43.436 --rc genhtml_legend=1 00:04:43.436 --rc geninfo_all_blocks=1 00:04:43.436 --rc geninfo_unexecuted_blocks=1 00:04:43.436 00:04:43.436 ' 00:04:43.436 16:43:04 -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:04:43.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:43.436 --rc genhtml_branch_coverage=1 00:04:43.436 --rc genhtml_function_coverage=1 00:04:43.436 --rc genhtml_legend=1 00:04:43.436 --rc geninfo_all_blocks=1 00:04:43.436 --rc geninfo_unexecuted_blocks=1 00:04:43.436 00:04:43.436 ' 00:04:43.436 16:43:04 -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:04:43.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:04:43.436 --rc genhtml_branch_coverage=1 00:04:43.436 --rc genhtml_function_coverage=1 00:04:43.436 --rc genhtml_legend=1 00:04:43.436 --rc geninfo_all_blocks=1 00:04:43.436 --rc geninfo_unexecuted_blocks=1 00:04:43.436 00:04:43.436 ' 00:04:43.436 16:43:04 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:04:43.436 16:43:04 -- nvmf/common.sh@7 -- # uname -s 00:04:43.436 16:43:05 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:43.436 16:43:05 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:43.436 16:43:05 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:43.436 16:43:05 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:43.436 16:43:05 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:43.436 16:43:05 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:43.436 16:43:05 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:43.436 16:43:05 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:43.436 16:43:05 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:43.436 16:43:05 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:43.436 16:43:05 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:015174e6-42cc-4d4f-853e-c7d76863366c 00:04:43.436 16:43:05 -- nvmf/common.sh@18 -- # NVME_HOSTID=015174e6-42cc-4d4f-853e-c7d76863366c 00:04:43.436 16:43:05 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:43.436 16:43:05 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:43.436 16:43:05 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:43.436 16:43:05 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:43.436 16:43:05 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:04:43.436 16:43:05 -- scripts/common.sh@15 -- # shopt -s extglob 00:04:43.436 16:43:05 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:43.436 16:43:05 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:43.436 16:43:05 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:43.436 16:43:05 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:43.436 16:43:05 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:43.436 16:43:05 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:43.436 16:43:05 -- paths/export.sh@5 -- # export PATH 00:04:43.436 16:43:05 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:43.436 16:43:05 -- nvmf/common.sh@51 -- # : 0 00:04:43.436 16:43:05 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:04:43.436 16:43:05 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:04:43.436 16:43:05 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:43.436 16:43:05 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:43.436 16:43:05 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:43.436 16:43:05 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:04:43.436 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:04:43.436 16:43:05 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:04:43.436 16:43:05 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:04:43.436 16:43:05 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:04:43.436 16:43:05 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:04:43.436 16:43:05 -- spdk/autotest.sh@32 -- # uname -s 00:04:43.436 16:43:05 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:04:43.436 16:43:05 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:04:43.436 16:43:05 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:43.436 16:43:05 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:04:43.436 16:43:05 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:04:43.436 16:43:05 -- spdk/autotest.sh@44 -- # modprobe nbd 00:04:43.436 16:43:05 -- spdk/autotest.sh@46 -- # type -P udevadm 00:04:43.436 16:43:05 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:04:43.436 16:43:05 -- spdk/autotest.sh@48 -- # udevadm_pid=66449 00:04:43.436 16:43:05 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:04:43.436 16:43:05 -- pm/common@17 -- # local monitor 00:04:43.436 16:43:05 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:43.436 16:43:05 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:04:43.696 16:43:05 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:04:43.696 16:43:05 -- pm/common@25 -- # sleep 1 00:04:43.696 16:43:05 -- pm/common@21 -- # date +%s 00:04:43.696 16:43:05 -- pm/common@21 -- # date +%s 00:04:43.696 16:43:05 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1727628185 00:04:43.696 16:43:05 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1727628185 00:04:43.696 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1727628185_collect-vmstat.pm.log 00:04:43.696 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1727628185_collect-cpu-load.pm.log 00:04:44.633 16:43:06 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:04:44.633 16:43:06 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:04:44.633 16:43:06 -- common/autotest_common.sh@724 -- # xtrace_disable 00:04:44.633 16:43:06 -- common/autotest_common.sh@10 -- # set +x 00:04:44.633 16:43:06 -- spdk/autotest.sh@59 -- # create_test_list 00:04:44.633 16:43:06 -- common/autotest_common.sh@748 -- # xtrace_disable 00:04:44.633 16:43:06 -- common/autotest_common.sh@10 -- # set +x 00:04:44.633 16:43:06 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:04:44.633 16:43:06 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:04:44.633 16:43:06 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:04:44.633 16:43:06 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:04:44.633 16:43:06 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:04:44.633 16:43:06 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:04:44.633 16:43:06 -- common/autotest_common.sh@1455 -- # uname 00:04:44.634 16:43:06 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:04:44.634 16:43:06 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:04:44.634 16:43:06 -- common/autotest_common.sh@1475 -- # uname 00:04:44.634 16:43:06 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:04:44.634 16:43:06 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:04:44.634 16:43:06 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:04:44.634 lcov: LCOV version 1.15 00:04:44.634 16:43:06 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:04:59.526 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:04:59.526 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:05:11.881 16:43:33 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:05:11.881 16:43:33 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:11.881 16:43:33 -- common/autotest_common.sh@10 -- # set +x 00:05:11.881 16:43:33 -- spdk/autotest.sh@78 -- # rm -f 00:05:11.881 16:43:33 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:12.451 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:12.710 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:05:12.710 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:05:12.710 16:43:34 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:05:12.710 16:43:34 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:05:12.710 16:43:34 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:05:12.711 16:43:34 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:05:12.711 16:43:34 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:12.711 16:43:34 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:05:12.711 16:43:34 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:05:12.711 16:43:34 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:12.711 16:43:34 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:12.711 16:43:34 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:12.711 16:43:34 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n2 00:05:12.711 16:43:34 -- common/autotest_common.sh@1648 -- # local device=nvme0n2 00:05:12.711 16:43:34 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n2/queue/zoned ]] 00:05:12.711 16:43:34 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:12.711 16:43:34 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:12.711 16:43:34 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n3 00:05:12.711 16:43:34 -- common/autotest_common.sh@1648 -- # local device=nvme0n3 00:05:12.711 16:43:34 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n3/queue/zoned ]] 00:05:12.711 16:43:34 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:12.711 16:43:34 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:12.711 16:43:34 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n1 00:05:12.711 16:43:34 -- common/autotest_common.sh@1648 -- # local device=nvme1n1 00:05:12.711 16:43:34 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:05:12.711 16:43:34 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:12.711 16:43:34 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:05:12.711 16:43:34 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:12.711 16:43:34 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:12.711 16:43:34 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:05:12.711 16:43:34 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:05:12.711 16:43:34 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:05:12.711 No valid GPT data, bailing 00:05:12.711 16:43:34 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:12.711 16:43:34 -- scripts/common.sh@394 -- # pt= 00:05:12.711 16:43:34 -- scripts/common.sh@395 -- # return 1 00:05:12.711 16:43:34 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:05:12.711 1+0 records in 00:05:12.711 1+0 records out 00:05:12.711 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00566767 s, 185 MB/s 00:05:12.711 16:43:34 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:12.711 16:43:34 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:12.711 16:43:34 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n2 00:05:12.711 16:43:34 -- scripts/common.sh@381 -- # local block=/dev/nvme0n2 pt 00:05:12.711 16:43:34 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n2 00:05:12.711 No valid GPT data, bailing 00:05:12.711 16:43:34 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n2 00:05:12.970 16:43:34 -- scripts/common.sh@394 -- # pt= 00:05:12.971 16:43:34 -- scripts/common.sh@395 -- # return 1 00:05:12.971 16:43:34 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n2 bs=1M count=1 00:05:12.971 1+0 records in 00:05:12.971 1+0 records out 00:05:12.971 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00458176 s, 229 MB/s 00:05:12.971 16:43:34 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:12.971 16:43:34 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:12.971 16:43:34 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n3 00:05:12.971 16:43:34 -- scripts/common.sh@381 -- # local block=/dev/nvme0n3 pt 00:05:12.971 16:43:34 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n3 00:05:12.971 No valid GPT data, bailing 00:05:12.971 16:43:34 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n3 00:05:12.971 16:43:34 -- scripts/common.sh@394 -- # pt= 00:05:12.971 16:43:34 -- scripts/common.sh@395 -- # return 1 00:05:12.971 16:43:34 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n3 bs=1M count=1 00:05:12.971 1+0 records in 00:05:12.971 1+0 records out 00:05:12.971 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00595666 s, 176 MB/s 00:05:12.971 16:43:34 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:12.971 16:43:34 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:12.971 16:43:34 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:05:12.971 16:43:34 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:05:12.971 16:43:34 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:05:12.971 No valid GPT data, bailing 00:05:12.971 16:43:34 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:05:12.971 16:43:34 -- scripts/common.sh@394 -- # pt= 00:05:12.971 16:43:34 -- scripts/common.sh@395 -- # return 1 00:05:12.971 16:43:34 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:05:12.971 1+0 records in 00:05:12.971 1+0 records out 00:05:12.971 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00645847 s, 162 MB/s 00:05:12.971 16:43:34 -- spdk/autotest.sh@105 -- # sync 00:05:13.540 16:43:34 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:05:13.540 16:43:34 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:05:13.540 16:43:34 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:05:16.836 16:43:37 -- spdk/autotest.sh@111 -- # uname -s 00:05:16.836 16:43:37 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:05:16.836 16:43:37 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:05:16.836 16:43:37 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:05:17.096 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:17.096 Hugepages 00:05:17.096 node hugesize free / total 00:05:17.096 node0 1048576kB 0 / 0 00:05:17.096 node0 2048kB 0 / 0 00:05:17.096 00:05:17.096 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:17.096 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:05:17.357 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme1 nvme1n1 00:05:17.357 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme0 nvme0n1 nvme0n2 nvme0n3 00:05:17.357 16:43:38 -- spdk/autotest.sh@117 -- # uname -s 00:05:17.357 16:43:38 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:05:17.357 16:43:38 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:05:17.357 16:43:38 -- common/autotest_common.sh@1514 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:18.302 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:18.302 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:18.302 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:18.302 16:43:39 -- common/autotest_common.sh@1515 -- # sleep 1 00:05:19.683 16:43:40 -- common/autotest_common.sh@1516 -- # bdfs=() 00:05:19.683 16:43:40 -- common/autotest_common.sh@1516 -- # local bdfs 00:05:19.683 16:43:40 -- common/autotest_common.sh@1518 -- # bdfs=($(get_nvme_bdfs)) 00:05:19.683 16:43:40 -- common/autotest_common.sh@1518 -- # get_nvme_bdfs 00:05:19.683 16:43:40 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:19.683 16:43:40 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:19.683 16:43:40 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:19.683 16:43:40 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:19.684 16:43:40 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:19.684 16:43:41 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:19.684 16:43:41 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:19.684 16:43:41 -- common/autotest_common.sh@1520 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:19.943 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:19.943 Waiting for block devices as requested 00:05:19.943 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:05:20.203 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:05:20.203 16:43:41 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:20.203 16:43:41 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1485 -- # grep 0000:00:10.0/nvme/nvme 00:05:20.203 16:43:41 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme1 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:20.203 16:43:41 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:20.203 16:43:41 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:20.203 16:43:41 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1541 -- # continue 00:05:20.203 16:43:41 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:20.203 16:43:41 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:20.203 16:43:41 -- common/autotest_common.sh@1485 -- # grep 0000:00:11.0/nvme/nvme 00:05:20.203 16:43:41 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme0 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:20.203 16:43:41 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:20.203 16:43:41 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:20.203 16:43:41 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme0 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:20.203 16:43:41 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:20.203 16:43:41 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:20.203 16:43:41 -- common/autotest_common.sh@1541 -- # continue 00:05:20.203 16:43:41 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:05:20.203 16:43:41 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:20.203 16:43:41 -- common/autotest_common.sh@10 -- # set +x 00:05:20.463 16:43:41 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:05:20.463 16:43:41 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:20.463 16:43:41 -- common/autotest_common.sh@10 -- # set +x 00:05:20.463 16:43:41 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:21.055 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:21.329 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:21.329 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:21.329 16:43:42 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:05:21.329 16:43:42 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:21.329 16:43:42 -- common/autotest_common.sh@10 -- # set +x 00:05:21.329 16:43:42 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:05:21.329 16:43:42 -- common/autotest_common.sh@1576 -- # mapfile -t bdfs 00:05:21.329 16:43:42 -- common/autotest_common.sh@1576 -- # get_nvme_bdfs_by_id 0x0a54 00:05:21.329 16:43:42 -- common/autotest_common.sh@1561 -- # bdfs=() 00:05:21.329 16:43:42 -- common/autotest_common.sh@1561 -- # _bdfs=() 00:05:21.329 16:43:42 -- common/autotest_common.sh@1561 -- # local bdfs _bdfs 00:05:21.329 16:43:42 -- common/autotest_common.sh@1562 -- # _bdfs=($(get_nvme_bdfs)) 00:05:21.329 16:43:42 -- common/autotest_common.sh@1562 -- # get_nvme_bdfs 00:05:21.329 16:43:42 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:21.329 16:43:42 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:21.329 16:43:42 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:21.329 16:43:42 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:21.329 16:43:42 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:21.589 16:43:43 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:21.589 16:43:43 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:21.589 16:43:43 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:21.589 16:43:43 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:05:21.589 16:43:43 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:21.589 16:43:43 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:21.589 16:43:43 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:21.589 16:43:43 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:05:21.589 16:43:43 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:21.589 16:43:43 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:21.589 16:43:43 -- common/autotest_common.sh@1570 -- # (( 0 > 0 )) 00:05:21.589 16:43:43 -- common/autotest_common.sh@1570 -- # return 0 00:05:21.589 16:43:43 -- common/autotest_common.sh@1577 -- # [[ -z '' ]] 00:05:21.589 16:43:43 -- common/autotest_common.sh@1578 -- # return 0 00:05:21.589 16:43:43 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:05:21.589 16:43:43 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:05:21.589 16:43:43 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:21.589 16:43:43 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:21.589 16:43:43 -- spdk/autotest.sh@149 -- # timing_enter lib 00:05:21.589 16:43:43 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:21.589 16:43:43 -- common/autotest_common.sh@10 -- # set +x 00:05:21.589 16:43:43 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:05:21.589 16:43:43 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:21.589 16:43:43 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:21.589 16:43:43 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:21.589 16:43:43 -- common/autotest_common.sh@10 -- # set +x 00:05:21.589 ************************************ 00:05:21.589 START TEST env 00:05:21.589 ************************************ 00:05:21.589 16:43:43 env -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:21.589 * Looking for test storage... 00:05:21.589 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:05:21.589 16:43:43 env -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:21.589 16:43:43 env -- common/autotest_common.sh@1681 -- # lcov --version 00:05:21.589 16:43:43 env -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:21.849 16:43:43 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:21.849 16:43:43 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:21.849 16:43:43 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:21.849 16:43:43 env -- scripts/common.sh@336 -- # IFS=.-: 00:05:21.849 16:43:43 env -- scripts/common.sh@336 -- # read -ra ver1 00:05:21.849 16:43:43 env -- scripts/common.sh@337 -- # IFS=.-: 00:05:21.849 16:43:43 env -- scripts/common.sh@337 -- # read -ra ver2 00:05:21.849 16:43:43 env -- scripts/common.sh@338 -- # local 'op=<' 00:05:21.849 16:43:43 env -- scripts/common.sh@340 -- # ver1_l=2 00:05:21.849 16:43:43 env -- scripts/common.sh@341 -- # ver2_l=1 00:05:21.849 16:43:43 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:21.849 16:43:43 env -- scripts/common.sh@344 -- # case "$op" in 00:05:21.849 16:43:43 env -- scripts/common.sh@345 -- # : 1 00:05:21.849 16:43:43 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:21.849 16:43:43 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:21.849 16:43:43 env -- scripts/common.sh@365 -- # decimal 1 00:05:21.849 16:43:43 env -- scripts/common.sh@353 -- # local d=1 00:05:21.849 16:43:43 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:21.849 16:43:43 env -- scripts/common.sh@355 -- # echo 1 00:05:21.849 16:43:43 env -- scripts/common.sh@365 -- # ver1[v]=1 00:05:21.849 16:43:43 env -- scripts/common.sh@366 -- # decimal 2 00:05:21.849 16:43:43 env -- scripts/common.sh@353 -- # local d=2 00:05:21.849 16:43:43 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:21.849 16:43:43 env -- scripts/common.sh@355 -- # echo 2 00:05:21.849 16:43:43 env -- scripts/common.sh@366 -- # ver2[v]=2 00:05:21.849 16:43:43 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:21.849 16:43:43 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:21.849 16:43:43 env -- scripts/common.sh@368 -- # return 0 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:21.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.849 --rc genhtml_branch_coverage=1 00:05:21.849 --rc genhtml_function_coverage=1 00:05:21.849 --rc genhtml_legend=1 00:05:21.849 --rc geninfo_all_blocks=1 00:05:21.849 --rc geninfo_unexecuted_blocks=1 00:05:21.849 00:05:21.849 ' 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:21.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.849 --rc genhtml_branch_coverage=1 00:05:21.849 --rc genhtml_function_coverage=1 00:05:21.849 --rc genhtml_legend=1 00:05:21.849 --rc geninfo_all_blocks=1 00:05:21.849 --rc geninfo_unexecuted_blocks=1 00:05:21.849 00:05:21.849 ' 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:21.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.849 --rc genhtml_branch_coverage=1 00:05:21.849 --rc genhtml_function_coverage=1 00:05:21.849 --rc genhtml_legend=1 00:05:21.849 --rc geninfo_all_blocks=1 00:05:21.849 --rc geninfo_unexecuted_blocks=1 00:05:21.849 00:05:21.849 ' 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:21.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:21.849 --rc genhtml_branch_coverage=1 00:05:21.849 --rc genhtml_function_coverage=1 00:05:21.849 --rc genhtml_legend=1 00:05:21.849 --rc geninfo_all_blocks=1 00:05:21.849 --rc geninfo_unexecuted_blocks=1 00:05:21.849 00:05:21.849 ' 00:05:21.849 16:43:43 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:21.849 16:43:43 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:21.849 16:43:43 env -- common/autotest_common.sh@10 -- # set +x 00:05:21.849 ************************************ 00:05:21.849 START TEST env_memory 00:05:21.849 ************************************ 00:05:21.849 16:43:43 env.env_memory -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:21.849 00:05:21.849 00:05:21.849 CUnit - A unit testing framework for C - Version 2.1-3 00:05:21.849 http://cunit.sourceforge.net/ 00:05:21.849 00:05:21.849 00:05:21.849 Suite: memory 00:05:21.849 Test: alloc and free memory map ...[2024-09-29 16:43:43.403586] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:05:21.849 passed 00:05:21.849 Test: mem map translation ...[2024-09-29 16:43:43.446113] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:05:21.850 [2024-09-29 16:43:43.446186] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:05:21.850 [2024-09-29 16:43:43.446287] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:05:21.850 [2024-09-29 16:43:43.446331] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:05:21.850 passed 00:05:21.850 Test: mem map registration ...[2024-09-29 16:43:43.508945] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:05:21.850 [2024-09-29 16:43:43.509024] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:05:22.110 passed 00:05:22.110 Test: mem map adjacent registrations ...passed 00:05:22.110 00:05:22.110 Run Summary: Type Total Ran Passed Failed Inactive 00:05:22.110 suites 1 1 n/a 0 0 00:05:22.110 tests 4 4 4 0 0 00:05:22.110 asserts 152 152 152 0 n/a 00:05:22.110 00:05:22.110 Elapsed time = 0.232 seconds 00:05:22.110 00:05:22.110 real 0m0.295s 00:05:22.110 user 0m0.252s 00:05:22.110 sys 0m0.029s 00:05:22.110 16:43:43 env.env_memory -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:22.110 16:43:43 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:05:22.110 ************************************ 00:05:22.110 END TEST env_memory 00:05:22.110 ************************************ 00:05:22.110 16:43:43 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:22.110 16:43:43 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:22.110 16:43:43 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:22.110 16:43:43 env -- common/autotest_common.sh@10 -- # set +x 00:05:22.110 ************************************ 00:05:22.110 START TEST env_vtophys 00:05:22.110 ************************************ 00:05:22.110 16:43:43 env.env_vtophys -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:22.110 EAL: lib.eal log level changed from notice to debug 00:05:22.110 EAL: Detected lcore 0 as core 0 on socket 0 00:05:22.110 EAL: Detected lcore 1 as core 0 on socket 0 00:05:22.110 EAL: Detected lcore 2 as core 0 on socket 0 00:05:22.111 EAL: Detected lcore 3 as core 0 on socket 0 00:05:22.111 EAL: Detected lcore 4 as core 0 on socket 0 00:05:22.111 EAL: Detected lcore 5 as core 0 on socket 0 00:05:22.111 EAL: Detected lcore 6 as core 0 on socket 0 00:05:22.111 EAL: Detected lcore 7 as core 0 on socket 0 00:05:22.111 EAL: Detected lcore 8 as core 0 on socket 0 00:05:22.111 EAL: Detected lcore 9 as core 0 on socket 0 00:05:22.111 EAL: Maximum logical cores by configuration: 128 00:05:22.111 EAL: Detected CPU lcores: 10 00:05:22.111 EAL: Detected NUMA nodes: 1 00:05:22.111 EAL: Checking presence of .so 'librte_eal.so.23.0' 00:05:22.111 EAL: Detected shared linkage of DPDK 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23.0 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23.0 00:05:22.111 EAL: Registered [vdev] bus. 00:05:22.111 EAL: bus.vdev log level changed from disabled to notice 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23.0 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23.0 00:05:22.111 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:05:22.111 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:05:22.111 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:05:22.111 EAL: No shared files mode enabled, IPC will be disabled 00:05:22.111 EAL: No shared files mode enabled, IPC is disabled 00:05:22.111 EAL: Selected IOVA mode 'PA' 00:05:22.111 EAL: Probing VFIO support... 00:05:22.111 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:22.111 EAL: VFIO modules not loaded, skipping VFIO support... 00:05:22.111 EAL: Ask a virtual area of 0x2e000 bytes 00:05:22.111 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:05:22.111 EAL: Setting up physically contiguous memory... 00:05:22.111 EAL: Setting maximum number of open files to 524288 00:05:22.111 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:05:22.111 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:05:22.111 EAL: Ask a virtual area of 0x61000 bytes 00:05:22.111 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:05:22.111 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:22.111 EAL: Ask a virtual area of 0x400000000 bytes 00:05:22.111 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:05:22.111 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:05:22.111 EAL: Ask a virtual area of 0x61000 bytes 00:05:22.111 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:05:22.111 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:22.111 EAL: Ask a virtual area of 0x400000000 bytes 00:05:22.111 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:05:22.111 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:05:22.111 EAL: Ask a virtual area of 0x61000 bytes 00:05:22.111 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:05:22.111 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:22.111 EAL: Ask a virtual area of 0x400000000 bytes 00:05:22.111 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:05:22.111 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:05:22.111 EAL: Ask a virtual area of 0x61000 bytes 00:05:22.111 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:05:22.111 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:22.111 EAL: Ask a virtual area of 0x400000000 bytes 00:05:22.111 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:05:22.111 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:05:22.111 EAL: Hugepages will be freed exactly as allocated. 00:05:22.111 EAL: No shared files mode enabled, IPC is disabled 00:05:22.111 EAL: No shared files mode enabled, IPC is disabled 00:05:22.371 EAL: TSC frequency is ~2290000 KHz 00:05:22.371 EAL: Main lcore 0 is ready (tid=7fb63fd05a40;cpuset=[0]) 00:05:22.371 EAL: Trying to obtain current memory policy. 00:05:22.371 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.371 EAL: Restoring previous memory policy: 0 00:05:22.371 EAL: request: mp_malloc_sync 00:05:22.371 EAL: No shared files mode enabled, IPC is disabled 00:05:22.371 EAL: Heap on socket 0 was expanded by 2MB 00:05:22.371 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:22.371 EAL: No shared files mode enabled, IPC is disabled 00:05:22.371 EAL: No PCI address specified using 'addr=' in: bus=pci 00:05:22.371 EAL: Mem event callback 'spdk:(nil)' registered 00:05:22.371 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:05:22.371 00:05:22.371 00:05:22.371 CUnit - A unit testing framework for C - Version 2.1-3 00:05:22.371 http://cunit.sourceforge.net/ 00:05:22.371 00:05:22.371 00:05:22.371 Suite: components_suite 00:05:22.631 Test: vtophys_malloc_test ...passed 00:05:22.631 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:05:22.631 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.631 EAL: Restoring previous memory policy: 4 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was expanded by 4MB 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was shrunk by 4MB 00:05:22.631 EAL: Trying to obtain current memory policy. 00:05:22.631 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.631 EAL: Restoring previous memory policy: 4 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was expanded by 6MB 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was shrunk by 6MB 00:05:22.631 EAL: Trying to obtain current memory policy. 00:05:22.631 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.631 EAL: Restoring previous memory policy: 4 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was expanded by 10MB 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was shrunk by 10MB 00:05:22.631 EAL: Trying to obtain current memory policy. 00:05:22.631 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.631 EAL: Restoring previous memory policy: 4 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was expanded by 18MB 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was shrunk by 18MB 00:05:22.631 EAL: Trying to obtain current memory policy. 00:05:22.631 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.631 EAL: Restoring previous memory policy: 4 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was expanded by 34MB 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was shrunk by 34MB 00:05:22.631 EAL: Trying to obtain current memory policy. 00:05:22.631 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.631 EAL: Restoring previous memory policy: 4 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was expanded by 66MB 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was shrunk by 66MB 00:05:22.631 EAL: Trying to obtain current memory policy. 00:05:22.631 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.631 EAL: Restoring previous memory policy: 4 00:05:22.631 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.631 EAL: request: mp_malloc_sync 00:05:22.631 EAL: No shared files mode enabled, IPC is disabled 00:05:22.631 EAL: Heap on socket 0 was expanded by 130MB 00:05:22.632 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.891 EAL: request: mp_malloc_sync 00:05:22.891 EAL: No shared files mode enabled, IPC is disabled 00:05:22.891 EAL: Heap on socket 0 was shrunk by 130MB 00:05:22.891 EAL: Trying to obtain current memory policy. 00:05:22.891 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.891 EAL: Restoring previous memory policy: 4 00:05:22.891 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.891 EAL: request: mp_malloc_sync 00:05:22.891 EAL: No shared files mode enabled, IPC is disabled 00:05:22.891 EAL: Heap on socket 0 was expanded by 258MB 00:05:22.891 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.891 EAL: request: mp_malloc_sync 00:05:22.891 EAL: No shared files mode enabled, IPC is disabled 00:05:22.891 EAL: Heap on socket 0 was shrunk by 258MB 00:05:22.891 EAL: Trying to obtain current memory policy. 00:05:22.891 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:22.891 EAL: Restoring previous memory policy: 4 00:05:22.891 EAL: Calling mem event callback 'spdk:(nil)' 00:05:22.891 EAL: request: mp_malloc_sync 00:05:22.891 EAL: No shared files mode enabled, IPC is disabled 00:05:22.891 EAL: Heap on socket 0 was expanded by 514MB 00:05:23.150 EAL: Calling mem event callback 'spdk:(nil)' 00:05:23.150 EAL: request: mp_malloc_sync 00:05:23.150 EAL: No shared files mode enabled, IPC is disabled 00:05:23.150 EAL: Heap on socket 0 was shrunk by 514MB 00:05:23.150 EAL: Trying to obtain current memory policy. 00:05:23.150 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:23.410 EAL: Restoring previous memory policy: 4 00:05:23.410 EAL: Calling mem event callback 'spdk:(nil)' 00:05:23.410 EAL: request: mp_malloc_sync 00:05:23.410 EAL: No shared files mode enabled, IPC is disabled 00:05:23.410 EAL: Heap on socket 0 was expanded by 1026MB 00:05:23.670 EAL: Calling mem event callback 'spdk:(nil)' 00:05:23.670 passed 00:05:23.670 00:05:23.670 Run Summary: Type Total Ran Passed Failed Inactive 00:05:23.670 suites 1 1 n/a 0 0 00:05:23.670 tests 2 2 2 0 0 00:05:23.670 asserts 5386 5386 5386 0 n/a 00:05:23.670 00:05:23.670 Elapsed time = 1.357 seconds 00:05:23.670 EAL: request: mp_malloc_sync 00:05:23.670 EAL: No shared files mode enabled, IPC is disabled 00:05:23.670 EAL: Heap on socket 0 was shrunk by 1026MB 00:05:23.670 EAL: Calling mem event callback 'spdk:(nil)' 00:05:23.670 EAL: request: mp_malloc_sync 00:05:23.670 EAL: No shared files mode enabled, IPC is disabled 00:05:23.670 EAL: Heap on socket 0 was shrunk by 2MB 00:05:23.670 EAL: No shared files mode enabled, IPC is disabled 00:05:23.670 EAL: No shared files mode enabled, IPC is disabled 00:05:23.670 EAL: No shared files mode enabled, IPC is disabled 00:05:23.670 00:05:23.670 real 0m1.595s 00:05:23.670 user 0m0.763s 00:05:23.670 sys 0m0.702s 00:05:23.670 16:43:45 env.env_vtophys -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:23.670 16:43:45 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:05:23.670 ************************************ 00:05:23.670 END TEST env_vtophys 00:05:23.670 ************************************ 00:05:23.670 16:43:45 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:23.670 16:43:45 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:23.670 16:43:45 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:23.670 16:43:45 env -- common/autotest_common.sh@10 -- # set +x 00:05:23.670 ************************************ 00:05:23.670 START TEST env_pci 00:05:23.670 ************************************ 00:05:23.670 16:43:45 env.env_pci -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:23.931 00:05:23.931 00:05:23.931 CUnit - A unit testing framework for C - Version 2.1-3 00:05:23.931 http://cunit.sourceforge.net/ 00:05:23.931 00:05:23.931 00:05:23.931 Suite: pci 00:05:23.931 Test: pci_hook ...[2024-09-29 16:43:45.369213] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1049:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 68685 has claimed it 00:05:23.931 passed 00:05:23.931 00:05:23.931 Run Summary: Type Total Ran Passed Failed Inactive 00:05:23.931 suites 1 1 n/a 0 0 00:05:23.931 tests 1 1 1 0 0 00:05:23.931 asserts 25 25 25 0 n/a 00:05:23.931 00:05:23.931 Elapsed time = 0.007 seconds 00:05:23.931 EAL: Cannot find device (10000:00:01.0) 00:05:23.931 EAL: Failed to attach device on primary process 00:05:23.931 00:05:23.931 real 0m0.092s 00:05:23.931 user 0m0.041s 00:05:23.931 sys 0m0.050s 00:05:23.931 16:43:45 env.env_pci -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:23.931 16:43:45 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:05:23.931 ************************************ 00:05:23.931 END TEST env_pci 00:05:23.931 ************************************ 00:05:23.931 16:43:45 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:05:23.931 16:43:45 env -- env/env.sh@15 -- # uname 00:05:23.931 16:43:45 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:05:23.931 16:43:45 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:05:23.931 16:43:45 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:23.931 16:43:45 env -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:05:23.931 16:43:45 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:23.931 16:43:45 env -- common/autotest_common.sh@10 -- # set +x 00:05:23.931 ************************************ 00:05:23.931 START TEST env_dpdk_post_init 00:05:23.931 ************************************ 00:05:23.931 16:43:45 env.env_dpdk_post_init -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:23.931 EAL: Detected CPU lcores: 10 00:05:23.931 EAL: Detected NUMA nodes: 1 00:05:23.931 EAL: Detected shared linkage of DPDK 00:05:23.931 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:23.931 EAL: Selected IOVA mode 'PA' 00:05:24.191 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:24.191 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:05:24.191 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:05:24.191 Starting DPDK initialization... 00:05:24.191 Starting SPDK post initialization... 00:05:24.191 SPDK NVMe probe 00:05:24.191 Attaching to 0000:00:10.0 00:05:24.191 Attaching to 0000:00:11.0 00:05:24.191 Attached to 0000:00:10.0 00:05:24.191 Attached to 0000:00:11.0 00:05:24.191 Cleaning up... 00:05:24.191 00:05:24.191 real 0m0.235s 00:05:24.191 user 0m0.065s 00:05:24.191 sys 0m0.072s 00:05:24.191 16:43:45 env.env_dpdk_post_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:24.191 16:43:45 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:05:24.191 ************************************ 00:05:24.191 END TEST env_dpdk_post_init 00:05:24.191 ************************************ 00:05:24.191 16:43:45 env -- env/env.sh@26 -- # uname 00:05:24.191 16:43:45 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:05:24.191 16:43:45 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:24.191 16:43:45 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:24.191 16:43:45 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:24.191 16:43:45 env -- common/autotest_common.sh@10 -- # set +x 00:05:24.191 ************************************ 00:05:24.191 START TEST env_mem_callbacks 00:05:24.191 ************************************ 00:05:24.191 16:43:45 env.env_mem_callbacks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:24.192 EAL: Detected CPU lcores: 10 00:05:24.192 EAL: Detected NUMA nodes: 1 00:05:24.192 EAL: Detected shared linkage of DPDK 00:05:24.192 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:24.192 EAL: Selected IOVA mode 'PA' 00:05:24.452 00:05:24.452 00:05:24.452 CUnit - A unit testing framework for C - Version 2.1-3 00:05:24.452 http://cunit.sourceforge.net/ 00:05:24.452 00:05:24.452 00:05:24.452 Suite: memory 00:05:24.452 Test: test ... 00:05:24.452 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:24.452 register 0x200000200000 2097152 00:05:24.452 malloc 3145728 00:05:24.452 register 0x200000400000 4194304 00:05:24.452 buf 0x200000500000 len 3145728 PASSED 00:05:24.452 malloc 64 00:05:24.452 buf 0x2000004fff40 len 64 PASSED 00:05:24.452 malloc 4194304 00:05:24.452 register 0x200000800000 6291456 00:05:24.452 buf 0x200000a00000 len 4194304 PASSED 00:05:24.452 free 0x200000500000 3145728 00:05:24.452 free 0x2000004fff40 64 00:05:24.452 unregister 0x200000400000 4194304 PASSED 00:05:24.452 free 0x200000a00000 4194304 00:05:24.452 unregister 0x200000800000 6291456 PASSED 00:05:24.452 malloc 8388608 00:05:24.452 register 0x200000400000 10485760 00:05:24.452 buf 0x200000600000 len 8388608 PASSED 00:05:24.452 free 0x200000600000 8388608 00:05:24.452 unregister 0x200000400000 10485760 PASSED 00:05:24.452 passed 00:05:24.452 00:05:24.452 Run Summary: Type Total Ran Passed Failed Inactive 00:05:24.452 suites 1 1 n/a 0 0 00:05:24.452 tests 1 1 1 0 0 00:05:24.452 asserts 15 15 15 0 n/a 00:05:24.452 00:05:24.452 Elapsed time = 0.013 seconds 00:05:24.452 00:05:24.452 real 0m0.183s 00:05:24.452 user 0m0.031s 00:05:24.452 sys 0m0.050s 00:05:24.452 16:43:45 env.env_mem_callbacks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:24.452 16:43:45 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:05:24.452 ************************************ 00:05:24.452 END TEST env_mem_callbacks 00:05:24.452 ************************************ 00:05:24.452 00:05:24.452 real 0m2.972s 00:05:24.452 user 0m1.373s 00:05:24.452 sys 0m1.267s 00:05:24.452 16:43:46 env -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:24.452 16:43:46 env -- common/autotest_common.sh@10 -- # set +x 00:05:24.452 ************************************ 00:05:24.452 END TEST env 00:05:24.452 ************************************ 00:05:24.452 16:43:46 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:24.452 16:43:46 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:24.452 16:43:46 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:24.452 16:43:46 -- common/autotest_common.sh@10 -- # set +x 00:05:24.452 ************************************ 00:05:24.452 START TEST rpc 00:05:24.452 ************************************ 00:05:24.452 16:43:46 rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:24.712 * Looking for test storage... 00:05:24.712 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:24.712 16:43:46 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:24.712 16:43:46 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:24.712 16:43:46 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:24.712 16:43:46 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:24.712 16:43:46 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:24.712 16:43:46 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:24.712 16:43:46 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:24.712 16:43:46 rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:24.712 16:43:46 rpc -- scripts/common.sh@345 -- # : 1 00:05:24.712 16:43:46 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:24.712 16:43:46 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:24.712 16:43:46 rpc -- scripts/common.sh@365 -- # decimal 1 00:05:24.712 16:43:46 rpc -- scripts/common.sh@353 -- # local d=1 00:05:24.712 16:43:46 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:24.712 16:43:46 rpc -- scripts/common.sh@355 -- # echo 1 00:05:24.712 16:43:46 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:24.712 16:43:46 rpc -- scripts/common.sh@366 -- # decimal 2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@353 -- # local d=2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:24.712 16:43:46 rpc -- scripts/common.sh@355 -- # echo 2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:24.712 16:43:46 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:24.712 16:43:46 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:24.712 16:43:46 rpc -- scripts/common.sh@368 -- # return 0 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:24.712 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.712 --rc genhtml_branch_coverage=1 00:05:24.712 --rc genhtml_function_coverage=1 00:05:24.712 --rc genhtml_legend=1 00:05:24.712 --rc geninfo_all_blocks=1 00:05:24.712 --rc geninfo_unexecuted_blocks=1 00:05:24.712 00:05:24.712 ' 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:24.712 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.712 --rc genhtml_branch_coverage=1 00:05:24.712 --rc genhtml_function_coverage=1 00:05:24.712 --rc genhtml_legend=1 00:05:24.712 --rc geninfo_all_blocks=1 00:05:24.712 --rc geninfo_unexecuted_blocks=1 00:05:24.712 00:05:24.712 ' 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:24.712 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.712 --rc genhtml_branch_coverage=1 00:05:24.712 --rc genhtml_function_coverage=1 00:05:24.712 --rc genhtml_legend=1 00:05:24.712 --rc geninfo_all_blocks=1 00:05:24.712 --rc geninfo_unexecuted_blocks=1 00:05:24.712 00:05:24.712 ' 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:24.712 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:24.712 --rc genhtml_branch_coverage=1 00:05:24.712 --rc genhtml_function_coverage=1 00:05:24.712 --rc genhtml_legend=1 00:05:24.712 --rc geninfo_all_blocks=1 00:05:24.712 --rc geninfo_unexecuted_blocks=1 00:05:24.712 00:05:24.712 ' 00:05:24.712 16:43:46 rpc -- rpc/rpc.sh@65 -- # spdk_pid=68812 00:05:24.712 16:43:46 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:05:24.712 16:43:46 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:24.712 16:43:46 rpc -- rpc/rpc.sh@67 -- # waitforlisten 68812 00:05:24.712 16:43:46 rpc -- common/autotest_common.sh@831 -- # '[' -z 68812 ']' 00:05:24.713 16:43:46 rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:24.713 16:43:46 rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:24.713 16:43:46 rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:24.713 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:24.713 16:43:46 rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:24.713 16:43:46 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:24.973 [2024-09-29 16:43:46.446036] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:24.973 [2024-09-29 16:43:46.446177] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68812 ] 00:05:24.973 [2024-09-29 16:43:46.591306] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:24.973 [2024-09-29 16:43:46.638035] app.c: 610:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:05:24.973 [2024-09-29 16:43:46.638103] app.c: 611:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 68812' to capture a snapshot of events at runtime. 00:05:24.973 [2024-09-29 16:43:46.638115] app.c: 616:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:24.973 [2024-09-29 16:43:46.638124] app.c: 617:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:24.973 [2024-09-29 16:43:46.638143] app.c: 618:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid68812 for offline analysis/debug. 00:05:24.973 [2024-09-29 16:43:46.638187] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:25.912 16:43:47 rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:25.912 16:43:47 rpc -- common/autotest_common.sh@864 -- # return 0 00:05:25.912 16:43:47 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:25.912 16:43:47 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:25.912 16:43:47 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:05:25.912 16:43:47 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:05:25.912 16:43:47 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:25.912 16:43:47 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:25.912 16:43:47 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:25.912 ************************************ 00:05:25.912 START TEST rpc_integrity 00:05:25.912 ************************************ 00:05:25.912 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:05:25.912 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:25.912 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.912 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.912 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.912 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:25.912 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:25.912 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:25.912 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:25.912 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.912 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:25.913 { 00:05:25.913 "name": "Malloc0", 00:05:25.913 "aliases": [ 00:05:25.913 "93f413f4-3add-428a-a742-aff4d89186e0" 00:05:25.913 ], 00:05:25.913 "product_name": "Malloc disk", 00:05:25.913 "block_size": 512, 00:05:25.913 "num_blocks": 16384, 00:05:25.913 "uuid": "93f413f4-3add-428a-a742-aff4d89186e0", 00:05:25.913 "assigned_rate_limits": { 00:05:25.913 "rw_ios_per_sec": 0, 00:05:25.913 "rw_mbytes_per_sec": 0, 00:05:25.913 "r_mbytes_per_sec": 0, 00:05:25.913 "w_mbytes_per_sec": 0 00:05:25.913 }, 00:05:25.913 "claimed": false, 00:05:25.913 "zoned": false, 00:05:25.913 "supported_io_types": { 00:05:25.913 "read": true, 00:05:25.913 "write": true, 00:05:25.913 "unmap": true, 00:05:25.913 "flush": true, 00:05:25.913 "reset": true, 00:05:25.913 "nvme_admin": false, 00:05:25.913 "nvme_io": false, 00:05:25.913 "nvme_io_md": false, 00:05:25.913 "write_zeroes": true, 00:05:25.913 "zcopy": true, 00:05:25.913 "get_zone_info": false, 00:05:25.913 "zone_management": false, 00:05:25.913 "zone_append": false, 00:05:25.913 "compare": false, 00:05:25.913 "compare_and_write": false, 00:05:25.913 "abort": true, 00:05:25.913 "seek_hole": false, 00:05:25.913 "seek_data": false, 00:05:25.913 "copy": true, 00:05:25.913 "nvme_iov_md": false 00:05:25.913 }, 00:05:25.913 "memory_domains": [ 00:05:25.913 { 00:05:25.913 "dma_device_id": "system", 00:05:25.913 "dma_device_type": 1 00:05:25.913 }, 00:05:25.913 { 00:05:25.913 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:25.913 "dma_device_type": 2 00:05:25.913 } 00:05:25.913 ], 00:05:25.913 "driver_specific": {} 00:05:25.913 } 00:05:25.913 ]' 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 [2024-09-29 16:43:47.409448] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:05:25.913 [2024-09-29 16:43:47.409503] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:25.913 [2024-09-29 16:43:47.409555] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006c80 00:05:25.913 [2024-09-29 16:43:47.409567] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:25.913 [2024-09-29 16:43:47.412055] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:25.913 [2024-09-29 16:43:47.412090] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:25.913 Passthru0 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:25.913 { 00:05:25.913 "name": "Malloc0", 00:05:25.913 "aliases": [ 00:05:25.913 "93f413f4-3add-428a-a742-aff4d89186e0" 00:05:25.913 ], 00:05:25.913 "product_name": "Malloc disk", 00:05:25.913 "block_size": 512, 00:05:25.913 "num_blocks": 16384, 00:05:25.913 "uuid": "93f413f4-3add-428a-a742-aff4d89186e0", 00:05:25.913 "assigned_rate_limits": { 00:05:25.913 "rw_ios_per_sec": 0, 00:05:25.913 "rw_mbytes_per_sec": 0, 00:05:25.913 "r_mbytes_per_sec": 0, 00:05:25.913 "w_mbytes_per_sec": 0 00:05:25.913 }, 00:05:25.913 "claimed": true, 00:05:25.913 "claim_type": "exclusive_write", 00:05:25.913 "zoned": false, 00:05:25.913 "supported_io_types": { 00:05:25.913 "read": true, 00:05:25.913 "write": true, 00:05:25.913 "unmap": true, 00:05:25.913 "flush": true, 00:05:25.913 "reset": true, 00:05:25.913 "nvme_admin": false, 00:05:25.913 "nvme_io": false, 00:05:25.913 "nvme_io_md": false, 00:05:25.913 "write_zeroes": true, 00:05:25.913 "zcopy": true, 00:05:25.913 "get_zone_info": false, 00:05:25.913 "zone_management": false, 00:05:25.913 "zone_append": false, 00:05:25.913 "compare": false, 00:05:25.913 "compare_and_write": false, 00:05:25.913 "abort": true, 00:05:25.913 "seek_hole": false, 00:05:25.913 "seek_data": false, 00:05:25.913 "copy": true, 00:05:25.913 "nvme_iov_md": false 00:05:25.913 }, 00:05:25.913 "memory_domains": [ 00:05:25.913 { 00:05:25.913 "dma_device_id": "system", 00:05:25.913 "dma_device_type": 1 00:05:25.913 }, 00:05:25.913 { 00:05:25.913 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:25.913 "dma_device_type": 2 00:05:25.913 } 00:05:25.913 ], 00:05:25.913 "driver_specific": {} 00:05:25.913 }, 00:05:25.913 { 00:05:25.913 "name": "Passthru0", 00:05:25.913 "aliases": [ 00:05:25.913 "f017ecc1-00cb-5bc9-bb6a-858705df975f" 00:05:25.913 ], 00:05:25.913 "product_name": "passthru", 00:05:25.913 "block_size": 512, 00:05:25.913 "num_blocks": 16384, 00:05:25.913 "uuid": "f017ecc1-00cb-5bc9-bb6a-858705df975f", 00:05:25.913 "assigned_rate_limits": { 00:05:25.913 "rw_ios_per_sec": 0, 00:05:25.913 "rw_mbytes_per_sec": 0, 00:05:25.913 "r_mbytes_per_sec": 0, 00:05:25.913 "w_mbytes_per_sec": 0 00:05:25.913 }, 00:05:25.913 "claimed": false, 00:05:25.913 "zoned": false, 00:05:25.913 "supported_io_types": { 00:05:25.913 "read": true, 00:05:25.913 "write": true, 00:05:25.913 "unmap": true, 00:05:25.913 "flush": true, 00:05:25.913 "reset": true, 00:05:25.913 "nvme_admin": false, 00:05:25.913 "nvme_io": false, 00:05:25.913 "nvme_io_md": false, 00:05:25.913 "write_zeroes": true, 00:05:25.913 "zcopy": true, 00:05:25.913 "get_zone_info": false, 00:05:25.913 "zone_management": false, 00:05:25.913 "zone_append": false, 00:05:25.913 "compare": false, 00:05:25.913 "compare_and_write": false, 00:05:25.913 "abort": true, 00:05:25.913 "seek_hole": false, 00:05:25.913 "seek_data": false, 00:05:25.913 "copy": true, 00:05:25.913 "nvme_iov_md": false 00:05:25.913 }, 00:05:25.913 "memory_domains": [ 00:05:25.913 { 00:05:25.913 "dma_device_id": "system", 00:05:25.913 "dma_device_type": 1 00:05:25.913 }, 00:05:25.913 { 00:05:25.913 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:25.913 "dma_device_type": 2 00:05:25.913 } 00:05:25.913 ], 00:05:25.913 "driver_specific": { 00:05:25.913 "passthru": { 00:05:25.913 "name": "Passthru0", 00:05:25.913 "base_bdev_name": "Malloc0" 00:05:25.913 } 00:05:25.913 } 00:05:25.913 } 00:05:25.913 ]' 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:25.913 16:43:47 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:25.913 00:05:25.913 real 0m0.292s 00:05:25.913 user 0m0.174s 00:05:25.913 sys 0m0.048s 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:25.913 16:43:47 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:25.913 ************************************ 00:05:25.913 END TEST rpc_integrity 00:05:25.913 ************************************ 00:05:26.173 16:43:47 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:05:26.173 16:43:47 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:26.173 16:43:47 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:26.173 16:43:47 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:26.173 ************************************ 00:05:26.173 START TEST rpc_plugins 00:05:26.173 ************************************ 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@1125 -- # rpc_plugins 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:05:26.173 { 00:05:26.173 "name": "Malloc1", 00:05:26.173 "aliases": [ 00:05:26.173 "6aa45258-f2ca-4204-880e-bbd8900953a8" 00:05:26.173 ], 00:05:26.173 "product_name": "Malloc disk", 00:05:26.173 "block_size": 4096, 00:05:26.173 "num_blocks": 256, 00:05:26.173 "uuid": "6aa45258-f2ca-4204-880e-bbd8900953a8", 00:05:26.173 "assigned_rate_limits": { 00:05:26.173 "rw_ios_per_sec": 0, 00:05:26.173 "rw_mbytes_per_sec": 0, 00:05:26.173 "r_mbytes_per_sec": 0, 00:05:26.173 "w_mbytes_per_sec": 0 00:05:26.173 }, 00:05:26.173 "claimed": false, 00:05:26.173 "zoned": false, 00:05:26.173 "supported_io_types": { 00:05:26.173 "read": true, 00:05:26.173 "write": true, 00:05:26.173 "unmap": true, 00:05:26.173 "flush": true, 00:05:26.173 "reset": true, 00:05:26.173 "nvme_admin": false, 00:05:26.173 "nvme_io": false, 00:05:26.173 "nvme_io_md": false, 00:05:26.173 "write_zeroes": true, 00:05:26.173 "zcopy": true, 00:05:26.173 "get_zone_info": false, 00:05:26.173 "zone_management": false, 00:05:26.173 "zone_append": false, 00:05:26.173 "compare": false, 00:05:26.173 "compare_and_write": false, 00:05:26.173 "abort": true, 00:05:26.173 "seek_hole": false, 00:05:26.173 "seek_data": false, 00:05:26.173 "copy": true, 00:05:26.173 "nvme_iov_md": false 00:05:26.173 }, 00:05:26.173 "memory_domains": [ 00:05:26.173 { 00:05:26.173 "dma_device_id": "system", 00:05:26.173 "dma_device_type": 1 00:05:26.173 }, 00:05:26.173 { 00:05:26.173 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:26.173 "dma_device_type": 2 00:05:26.173 } 00:05:26.173 ], 00:05:26.173 "driver_specific": {} 00:05:26.173 } 00:05:26.173 ]' 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:05:26.173 16:43:47 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:05:26.173 00:05:26.173 real 0m0.162s 00:05:26.173 user 0m0.091s 00:05:26.173 sys 0m0.029s 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:26.173 16:43:47 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:26.173 ************************************ 00:05:26.173 END TEST rpc_plugins 00:05:26.173 ************************************ 00:05:26.173 16:43:47 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:05:26.173 16:43:47 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:26.173 16:43:47 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:26.173 16:43:47 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:26.173 ************************************ 00:05:26.173 START TEST rpc_trace_cmd_test 00:05:26.173 ************************************ 00:05:26.173 16:43:47 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1125 -- # rpc_trace_cmd_test 00:05:26.173 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:05:26.433 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid68812", 00:05:26.433 "tpoint_group_mask": "0x8", 00:05:26.433 "iscsi_conn": { 00:05:26.433 "mask": "0x2", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "scsi": { 00:05:26.433 "mask": "0x4", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "bdev": { 00:05:26.433 "mask": "0x8", 00:05:26.433 "tpoint_mask": "0xffffffffffffffff" 00:05:26.433 }, 00:05:26.433 "nvmf_rdma": { 00:05:26.433 "mask": "0x10", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "nvmf_tcp": { 00:05:26.433 "mask": "0x20", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "ftl": { 00:05:26.433 "mask": "0x40", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "blobfs": { 00:05:26.433 "mask": "0x80", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "dsa": { 00:05:26.433 "mask": "0x200", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "thread": { 00:05:26.433 "mask": "0x400", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "nvme_pcie": { 00:05:26.433 "mask": "0x800", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "iaa": { 00:05:26.433 "mask": "0x1000", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "nvme_tcp": { 00:05:26.433 "mask": "0x2000", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "bdev_nvme": { 00:05:26.433 "mask": "0x4000", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "sock": { 00:05:26.433 "mask": "0x8000", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "blob": { 00:05:26.433 "mask": "0x10000", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 }, 00:05:26.433 "bdev_raid": { 00:05:26.433 "mask": "0x20000", 00:05:26.433 "tpoint_mask": "0x0" 00:05:26.433 } 00:05:26.433 }' 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 18 -gt 2 ']' 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:05:26.433 16:43:47 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:05:26.433 16:43:48 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:05:26.433 16:43:48 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:05:26.433 16:43:48 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:05:26.433 00:05:26.433 real 0m0.231s 00:05:26.433 user 0m0.183s 00:05:26.433 sys 0m0.034s 00:05:26.433 16:43:48 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:26.433 16:43:48 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:26.433 ************************************ 00:05:26.433 END TEST rpc_trace_cmd_test 00:05:26.433 ************************************ 00:05:26.693 16:43:48 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:05:26.693 16:43:48 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:05:26.693 16:43:48 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:05:26.693 16:43:48 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:26.693 16:43:48 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:26.693 16:43:48 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:26.693 ************************************ 00:05:26.693 START TEST rpc_daemon_integrity 00:05:26.693 ************************************ 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:26.693 { 00:05:26.693 "name": "Malloc2", 00:05:26.693 "aliases": [ 00:05:26.693 "5f0e3e7b-3d7c-4cda-88ca-7e4d8f2ad80e" 00:05:26.693 ], 00:05:26.693 "product_name": "Malloc disk", 00:05:26.693 "block_size": 512, 00:05:26.693 "num_blocks": 16384, 00:05:26.693 "uuid": "5f0e3e7b-3d7c-4cda-88ca-7e4d8f2ad80e", 00:05:26.693 "assigned_rate_limits": { 00:05:26.693 "rw_ios_per_sec": 0, 00:05:26.693 "rw_mbytes_per_sec": 0, 00:05:26.693 "r_mbytes_per_sec": 0, 00:05:26.693 "w_mbytes_per_sec": 0 00:05:26.693 }, 00:05:26.693 "claimed": false, 00:05:26.693 "zoned": false, 00:05:26.693 "supported_io_types": { 00:05:26.693 "read": true, 00:05:26.693 "write": true, 00:05:26.693 "unmap": true, 00:05:26.693 "flush": true, 00:05:26.693 "reset": true, 00:05:26.693 "nvme_admin": false, 00:05:26.693 "nvme_io": false, 00:05:26.693 "nvme_io_md": false, 00:05:26.693 "write_zeroes": true, 00:05:26.693 "zcopy": true, 00:05:26.693 "get_zone_info": false, 00:05:26.693 "zone_management": false, 00:05:26.693 "zone_append": false, 00:05:26.693 "compare": false, 00:05:26.693 "compare_and_write": false, 00:05:26.693 "abort": true, 00:05:26.693 "seek_hole": false, 00:05:26.693 "seek_data": false, 00:05:26.693 "copy": true, 00:05:26.693 "nvme_iov_md": false 00:05:26.693 }, 00:05:26.693 "memory_domains": [ 00:05:26.693 { 00:05:26.693 "dma_device_id": "system", 00:05:26.693 "dma_device_type": 1 00:05:26.693 }, 00:05:26.693 { 00:05:26.693 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:26.693 "dma_device_type": 2 00:05:26.693 } 00:05:26.693 ], 00:05:26.693 "driver_specific": {} 00:05:26.693 } 00:05:26.693 ]' 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.693 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.693 [2024-09-29 16:43:48.280597] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:05:26.693 [2024-09-29 16:43:48.280658] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:26.693 [2024-09-29 16:43:48.280683] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:05:26.693 [2024-09-29 16:43:48.280692] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:26.694 [2024-09-29 16:43:48.283033] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:26.694 [2024-09-29 16:43:48.283069] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:26.694 Passthru0 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:26.694 { 00:05:26.694 "name": "Malloc2", 00:05:26.694 "aliases": [ 00:05:26.694 "5f0e3e7b-3d7c-4cda-88ca-7e4d8f2ad80e" 00:05:26.694 ], 00:05:26.694 "product_name": "Malloc disk", 00:05:26.694 "block_size": 512, 00:05:26.694 "num_blocks": 16384, 00:05:26.694 "uuid": "5f0e3e7b-3d7c-4cda-88ca-7e4d8f2ad80e", 00:05:26.694 "assigned_rate_limits": { 00:05:26.694 "rw_ios_per_sec": 0, 00:05:26.694 "rw_mbytes_per_sec": 0, 00:05:26.694 "r_mbytes_per_sec": 0, 00:05:26.694 "w_mbytes_per_sec": 0 00:05:26.694 }, 00:05:26.694 "claimed": true, 00:05:26.694 "claim_type": "exclusive_write", 00:05:26.694 "zoned": false, 00:05:26.694 "supported_io_types": { 00:05:26.694 "read": true, 00:05:26.694 "write": true, 00:05:26.694 "unmap": true, 00:05:26.694 "flush": true, 00:05:26.694 "reset": true, 00:05:26.694 "nvme_admin": false, 00:05:26.694 "nvme_io": false, 00:05:26.694 "nvme_io_md": false, 00:05:26.694 "write_zeroes": true, 00:05:26.694 "zcopy": true, 00:05:26.694 "get_zone_info": false, 00:05:26.694 "zone_management": false, 00:05:26.694 "zone_append": false, 00:05:26.694 "compare": false, 00:05:26.694 "compare_and_write": false, 00:05:26.694 "abort": true, 00:05:26.694 "seek_hole": false, 00:05:26.694 "seek_data": false, 00:05:26.694 "copy": true, 00:05:26.694 "nvme_iov_md": false 00:05:26.694 }, 00:05:26.694 "memory_domains": [ 00:05:26.694 { 00:05:26.694 "dma_device_id": "system", 00:05:26.694 "dma_device_type": 1 00:05:26.694 }, 00:05:26.694 { 00:05:26.694 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:26.694 "dma_device_type": 2 00:05:26.694 } 00:05:26.694 ], 00:05:26.694 "driver_specific": {} 00:05:26.694 }, 00:05:26.694 { 00:05:26.694 "name": "Passthru0", 00:05:26.694 "aliases": [ 00:05:26.694 "a58c3950-be65-573c-8e38-4744355b76c5" 00:05:26.694 ], 00:05:26.694 "product_name": "passthru", 00:05:26.694 "block_size": 512, 00:05:26.694 "num_blocks": 16384, 00:05:26.694 "uuid": "a58c3950-be65-573c-8e38-4744355b76c5", 00:05:26.694 "assigned_rate_limits": { 00:05:26.694 "rw_ios_per_sec": 0, 00:05:26.694 "rw_mbytes_per_sec": 0, 00:05:26.694 "r_mbytes_per_sec": 0, 00:05:26.694 "w_mbytes_per_sec": 0 00:05:26.694 }, 00:05:26.694 "claimed": false, 00:05:26.694 "zoned": false, 00:05:26.694 "supported_io_types": { 00:05:26.694 "read": true, 00:05:26.694 "write": true, 00:05:26.694 "unmap": true, 00:05:26.694 "flush": true, 00:05:26.694 "reset": true, 00:05:26.694 "nvme_admin": false, 00:05:26.694 "nvme_io": false, 00:05:26.694 "nvme_io_md": false, 00:05:26.694 "write_zeroes": true, 00:05:26.694 "zcopy": true, 00:05:26.694 "get_zone_info": false, 00:05:26.694 "zone_management": false, 00:05:26.694 "zone_append": false, 00:05:26.694 "compare": false, 00:05:26.694 "compare_and_write": false, 00:05:26.694 "abort": true, 00:05:26.694 "seek_hole": false, 00:05:26.694 "seek_data": false, 00:05:26.694 "copy": true, 00:05:26.694 "nvme_iov_md": false 00:05:26.694 }, 00:05:26.694 "memory_domains": [ 00:05:26.694 { 00:05:26.694 "dma_device_id": "system", 00:05:26.694 "dma_device_type": 1 00:05:26.694 }, 00:05:26.694 { 00:05:26.694 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:26.694 "dma_device_type": 2 00:05:26.694 } 00:05:26.694 ], 00:05:26.694 "driver_specific": { 00:05:26.694 "passthru": { 00:05:26.694 "name": "Passthru0", 00:05:26.694 "base_bdev_name": "Malloc2" 00:05:26.694 } 00:05:26.694 } 00:05:26.694 } 00:05:26.694 ]' 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.694 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:26.954 00:05:26.954 real 0m0.304s 00:05:26.954 user 0m0.182s 00:05:26.954 sys 0m0.055s 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:26.954 16:43:48 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:26.954 ************************************ 00:05:26.954 END TEST rpc_daemon_integrity 00:05:26.954 ************************************ 00:05:26.954 16:43:48 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:05:26.954 16:43:48 rpc -- rpc/rpc.sh@84 -- # killprocess 68812 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@950 -- # '[' -z 68812 ']' 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@954 -- # kill -0 68812 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@955 -- # uname 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 68812 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:26.954 killing process with pid 68812 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 68812' 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@969 -- # kill 68812 00:05:26.954 16:43:48 rpc -- common/autotest_common.sh@974 -- # wait 68812 00:05:27.522 00:05:27.522 real 0m2.814s 00:05:27.522 user 0m3.290s 00:05:27.522 sys 0m0.897s 00:05:27.522 16:43:48 rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:27.522 16:43:48 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:27.522 ************************************ 00:05:27.522 END TEST rpc 00:05:27.522 ************************************ 00:05:27.522 16:43:48 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:27.522 16:43:48 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:27.522 16:43:48 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:27.522 16:43:48 -- common/autotest_common.sh@10 -- # set +x 00:05:27.522 ************************************ 00:05:27.522 START TEST skip_rpc 00:05:27.522 ************************************ 00:05:27.522 16:43:48 skip_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:27.522 * Looking for test storage... 00:05:27.522 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:27.522 16:43:49 skip_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:27.522 16:43:49 skip_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:27.522 16:43:49 skip_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:27.522 16:43:49 skip_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@345 -- # : 1 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:27.522 16:43:49 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:27.782 16:43:49 skip_rpc -- scripts/common.sh@368 -- # return 0 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:27.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.782 --rc genhtml_branch_coverage=1 00:05:27.782 --rc genhtml_function_coverage=1 00:05:27.782 --rc genhtml_legend=1 00:05:27.782 --rc geninfo_all_blocks=1 00:05:27.782 --rc geninfo_unexecuted_blocks=1 00:05:27.782 00:05:27.782 ' 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:27.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.782 --rc genhtml_branch_coverage=1 00:05:27.782 --rc genhtml_function_coverage=1 00:05:27.782 --rc genhtml_legend=1 00:05:27.782 --rc geninfo_all_blocks=1 00:05:27.782 --rc geninfo_unexecuted_blocks=1 00:05:27.782 00:05:27.782 ' 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:27.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.782 --rc genhtml_branch_coverage=1 00:05:27.782 --rc genhtml_function_coverage=1 00:05:27.782 --rc genhtml_legend=1 00:05:27.782 --rc geninfo_all_blocks=1 00:05:27.782 --rc geninfo_unexecuted_blocks=1 00:05:27.782 00:05:27.782 ' 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:27.782 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:27.782 --rc genhtml_branch_coverage=1 00:05:27.782 --rc genhtml_function_coverage=1 00:05:27.782 --rc genhtml_legend=1 00:05:27.782 --rc geninfo_all_blocks=1 00:05:27.782 --rc geninfo_unexecuted_blocks=1 00:05:27.782 00:05:27.782 ' 00:05:27.782 16:43:49 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:27.782 16:43:49 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:27.782 16:43:49 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:27.782 16:43:49 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:27.782 ************************************ 00:05:27.782 START TEST skip_rpc 00:05:27.782 ************************************ 00:05:27.782 16:43:49 skip_rpc.skip_rpc -- common/autotest_common.sh@1125 -- # test_skip_rpc 00:05:27.782 16:43:49 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69013 00:05:27.782 16:43:49 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:05:27.782 16:43:49 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:27.782 16:43:49 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:05:27.782 [2024-09-29 16:43:49.320186] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:27.782 [2024-09-29 16:43:49.320308] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69013 ] 00:05:28.041 [2024-09-29 16:43:49.466245] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:28.041 [2024-09-29 16:43:49.512030] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # local es=0 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd spdk_get_version 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # rpc_cmd spdk_get_version 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # es=1 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69013 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@950 -- # '[' -z 69013 ']' 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # kill -0 69013 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # uname 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69013 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69013' 00:05:33.314 killing process with pid 69013 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@969 -- # kill 69013 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@974 -- # wait 69013 00:05:33.314 00:05:33.314 real 0m5.446s 00:05:33.314 user 0m5.050s 00:05:33.314 sys 0m0.319s 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:33.314 16:43:54 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.314 ************************************ 00:05:33.314 END TEST skip_rpc 00:05:33.314 ************************************ 00:05:33.314 16:43:54 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:05:33.314 16:43:54 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:33.314 16:43:54 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:33.314 16:43:54 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.314 ************************************ 00:05:33.314 START TEST skip_rpc_with_json 00:05:33.314 ************************************ 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_json 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69101 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69101 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@831 -- # '[' -z 69101 ']' 00:05:33.314 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:33.314 16:43:54 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:33.314 [2024-09-29 16:43:54.827203] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:33.314 [2024-09-29 16:43:54.827332] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69101 ] 00:05:33.314 [2024-09-29 16:43:54.970866] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:33.573 [2024-09-29 16:43:55.017209] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:34.140 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:34.140 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # return 0 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:34.141 [2024-09-29 16:43:55.641394] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:05:34.141 request: 00:05:34.141 { 00:05:34.141 "trtype": "tcp", 00:05:34.141 "method": "nvmf_get_transports", 00:05:34.141 "req_id": 1 00:05:34.141 } 00:05:34.141 Got JSON-RPC error response 00:05:34.141 response: 00:05:34.141 { 00:05:34.141 "code": -19, 00:05:34.141 "message": "No such device" 00:05:34.141 } 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:34.141 [2024-09-29 16:43:55.653466] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:34.141 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:34.400 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:34.400 16:43:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:34.400 { 00:05:34.400 "subsystems": [ 00:05:34.400 { 00:05:34.400 "subsystem": "fsdev", 00:05:34.400 "config": [ 00:05:34.400 { 00:05:34.400 "method": "fsdev_set_opts", 00:05:34.400 "params": { 00:05:34.400 "fsdev_io_pool_size": 65535, 00:05:34.400 "fsdev_io_cache_size": 256 00:05:34.400 } 00:05:34.400 } 00:05:34.400 ] 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "subsystem": "keyring", 00:05:34.400 "config": [] 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "subsystem": "iobuf", 00:05:34.400 "config": [ 00:05:34.400 { 00:05:34.400 "method": "iobuf_set_options", 00:05:34.400 "params": { 00:05:34.400 "small_pool_count": 8192, 00:05:34.400 "large_pool_count": 1024, 00:05:34.400 "small_bufsize": 8192, 00:05:34.400 "large_bufsize": 135168 00:05:34.400 } 00:05:34.400 } 00:05:34.400 ] 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "subsystem": "sock", 00:05:34.400 "config": [ 00:05:34.400 { 00:05:34.400 "method": "sock_set_default_impl", 00:05:34.400 "params": { 00:05:34.400 "impl_name": "posix" 00:05:34.400 } 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "method": "sock_impl_set_options", 00:05:34.400 "params": { 00:05:34.400 "impl_name": "ssl", 00:05:34.400 "recv_buf_size": 4096, 00:05:34.400 "send_buf_size": 4096, 00:05:34.400 "enable_recv_pipe": true, 00:05:34.400 "enable_quickack": false, 00:05:34.400 "enable_placement_id": 0, 00:05:34.400 "enable_zerocopy_send_server": true, 00:05:34.400 "enable_zerocopy_send_client": false, 00:05:34.400 "zerocopy_threshold": 0, 00:05:34.400 "tls_version": 0, 00:05:34.400 "enable_ktls": false 00:05:34.400 } 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "method": "sock_impl_set_options", 00:05:34.400 "params": { 00:05:34.400 "impl_name": "posix", 00:05:34.400 "recv_buf_size": 2097152, 00:05:34.400 "send_buf_size": 2097152, 00:05:34.400 "enable_recv_pipe": true, 00:05:34.400 "enable_quickack": false, 00:05:34.400 "enable_placement_id": 0, 00:05:34.400 "enable_zerocopy_send_server": true, 00:05:34.400 "enable_zerocopy_send_client": false, 00:05:34.400 "zerocopy_threshold": 0, 00:05:34.400 "tls_version": 0, 00:05:34.400 "enable_ktls": false 00:05:34.400 } 00:05:34.400 } 00:05:34.400 ] 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "subsystem": "vmd", 00:05:34.400 "config": [] 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "subsystem": "accel", 00:05:34.400 "config": [ 00:05:34.400 { 00:05:34.400 "method": "accel_set_options", 00:05:34.400 "params": { 00:05:34.400 "small_cache_size": 128, 00:05:34.400 "large_cache_size": 16, 00:05:34.400 "task_count": 2048, 00:05:34.400 "sequence_count": 2048, 00:05:34.400 "buf_count": 2048 00:05:34.400 } 00:05:34.400 } 00:05:34.400 ] 00:05:34.400 }, 00:05:34.400 { 00:05:34.400 "subsystem": "bdev", 00:05:34.400 "config": [ 00:05:34.400 { 00:05:34.400 "method": "bdev_set_options", 00:05:34.400 "params": { 00:05:34.400 "bdev_io_pool_size": 65535, 00:05:34.400 "bdev_io_cache_size": 256, 00:05:34.400 "bdev_auto_examine": true, 00:05:34.400 "iobuf_small_cache_size": 128, 00:05:34.400 "iobuf_large_cache_size": 16 00:05:34.400 } 00:05:34.400 }, 00:05:34.400 { 00:05:34.401 "method": "bdev_raid_set_options", 00:05:34.401 "params": { 00:05:34.401 "process_window_size_kb": 1024, 00:05:34.401 "process_max_bandwidth_mb_sec": 0 00:05:34.401 } 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "method": "bdev_iscsi_set_options", 00:05:34.401 "params": { 00:05:34.401 "timeout_sec": 30 00:05:34.401 } 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "method": "bdev_nvme_set_options", 00:05:34.401 "params": { 00:05:34.401 "action_on_timeout": "none", 00:05:34.401 "timeout_us": 0, 00:05:34.401 "timeout_admin_us": 0, 00:05:34.401 "keep_alive_timeout_ms": 10000, 00:05:34.401 "arbitration_burst": 0, 00:05:34.401 "low_priority_weight": 0, 00:05:34.401 "medium_priority_weight": 0, 00:05:34.401 "high_priority_weight": 0, 00:05:34.401 "nvme_adminq_poll_period_us": 10000, 00:05:34.401 "nvme_ioq_poll_period_us": 0, 00:05:34.401 "io_queue_requests": 0, 00:05:34.401 "delay_cmd_submit": true, 00:05:34.401 "transport_retry_count": 4, 00:05:34.401 "bdev_retry_count": 3, 00:05:34.401 "transport_ack_timeout": 0, 00:05:34.401 "ctrlr_loss_timeout_sec": 0, 00:05:34.401 "reconnect_delay_sec": 0, 00:05:34.401 "fast_io_fail_timeout_sec": 0, 00:05:34.401 "disable_auto_failback": false, 00:05:34.401 "generate_uuids": false, 00:05:34.401 "transport_tos": 0, 00:05:34.401 "nvme_error_stat": false, 00:05:34.401 "rdma_srq_size": 0, 00:05:34.401 "io_path_stat": false, 00:05:34.401 "allow_accel_sequence": false, 00:05:34.401 "rdma_max_cq_size": 0, 00:05:34.401 "rdma_cm_event_timeout_ms": 0, 00:05:34.401 "dhchap_digests": [ 00:05:34.401 "sha256", 00:05:34.401 "sha384", 00:05:34.401 "sha512" 00:05:34.401 ], 00:05:34.401 "dhchap_dhgroups": [ 00:05:34.401 "null", 00:05:34.401 "ffdhe2048", 00:05:34.401 "ffdhe3072", 00:05:34.401 "ffdhe4096", 00:05:34.401 "ffdhe6144", 00:05:34.401 "ffdhe8192" 00:05:34.401 ] 00:05:34.401 } 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "method": "bdev_nvme_set_hotplug", 00:05:34.401 "params": { 00:05:34.401 "period_us": 100000, 00:05:34.401 "enable": false 00:05:34.401 } 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "method": "bdev_wait_for_examine" 00:05:34.401 } 00:05:34.401 ] 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "scsi", 00:05:34.401 "config": null 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "scheduler", 00:05:34.401 "config": [ 00:05:34.401 { 00:05:34.401 "method": "framework_set_scheduler", 00:05:34.401 "params": { 00:05:34.401 "name": "static" 00:05:34.401 } 00:05:34.401 } 00:05:34.401 ] 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "vhost_scsi", 00:05:34.401 "config": [] 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "vhost_blk", 00:05:34.401 "config": [] 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "ublk", 00:05:34.401 "config": [] 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "nbd", 00:05:34.401 "config": [] 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "nvmf", 00:05:34.401 "config": [ 00:05:34.401 { 00:05:34.401 "method": "nvmf_set_config", 00:05:34.401 "params": { 00:05:34.401 "discovery_filter": "match_any", 00:05:34.401 "admin_cmd_passthru": { 00:05:34.401 "identify_ctrlr": false 00:05:34.401 }, 00:05:34.401 "dhchap_digests": [ 00:05:34.401 "sha256", 00:05:34.401 "sha384", 00:05:34.401 "sha512" 00:05:34.401 ], 00:05:34.401 "dhchap_dhgroups": [ 00:05:34.401 "null", 00:05:34.401 "ffdhe2048", 00:05:34.401 "ffdhe3072", 00:05:34.401 "ffdhe4096", 00:05:34.401 "ffdhe6144", 00:05:34.401 "ffdhe8192" 00:05:34.401 ] 00:05:34.401 } 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "method": "nvmf_set_max_subsystems", 00:05:34.401 "params": { 00:05:34.401 "max_subsystems": 1024 00:05:34.401 } 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "method": "nvmf_set_crdt", 00:05:34.401 "params": { 00:05:34.401 "crdt1": 0, 00:05:34.401 "crdt2": 0, 00:05:34.401 "crdt3": 0 00:05:34.401 } 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "method": "nvmf_create_transport", 00:05:34.401 "params": { 00:05:34.401 "trtype": "TCP", 00:05:34.401 "max_queue_depth": 128, 00:05:34.401 "max_io_qpairs_per_ctrlr": 127, 00:05:34.401 "in_capsule_data_size": 4096, 00:05:34.401 "max_io_size": 131072, 00:05:34.401 "io_unit_size": 131072, 00:05:34.401 "max_aq_depth": 128, 00:05:34.401 "num_shared_buffers": 511, 00:05:34.401 "buf_cache_size": 4294967295, 00:05:34.401 "dif_insert_or_strip": false, 00:05:34.401 "zcopy": false, 00:05:34.401 "c2h_success": true, 00:05:34.401 "sock_priority": 0, 00:05:34.401 "abort_timeout_sec": 1, 00:05:34.401 "ack_timeout": 0, 00:05:34.401 "data_wr_pool_size": 0 00:05:34.401 } 00:05:34.401 } 00:05:34.401 ] 00:05:34.401 }, 00:05:34.401 { 00:05:34.401 "subsystem": "iscsi", 00:05:34.401 "config": [ 00:05:34.401 { 00:05:34.401 "method": "iscsi_set_options", 00:05:34.401 "params": { 00:05:34.401 "node_base": "iqn.2016-06.io.spdk", 00:05:34.401 "max_sessions": 128, 00:05:34.401 "max_connections_per_session": 2, 00:05:34.401 "max_queue_depth": 64, 00:05:34.401 "default_time2wait": 2, 00:05:34.401 "default_time2retain": 20, 00:05:34.401 "first_burst_length": 8192, 00:05:34.401 "immediate_data": true, 00:05:34.401 "allow_duplicated_isid": false, 00:05:34.401 "error_recovery_level": 0, 00:05:34.401 "nop_timeout": 60, 00:05:34.401 "nop_in_interval": 30, 00:05:34.401 "disable_chap": false, 00:05:34.401 "require_chap": false, 00:05:34.401 "mutual_chap": false, 00:05:34.401 "chap_group": 0, 00:05:34.401 "max_large_datain_per_connection": 64, 00:05:34.401 "max_r2t_per_connection": 4, 00:05:34.401 "pdu_pool_size": 36864, 00:05:34.401 "immediate_data_pool_size": 16384, 00:05:34.401 "data_out_pool_size": 2048 00:05:34.401 } 00:05:34.401 } 00:05:34.401 ] 00:05:34.401 } 00:05:34.401 ] 00:05:34.401 } 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69101 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 69101 ']' 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 69101 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69101 00:05:34.401 killing process with pid 69101 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69101' 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 69101 00:05:34.401 16:43:55 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 69101 00:05:34.660 16:43:56 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69124 00:05:34.660 16:43:56 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:34.660 16:43:56 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69124 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 69124 ']' 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 69124 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69124 00:05:39.932 killing process with pid 69124 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69124' 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 69124 00:05:39.932 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 69124 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:40.192 00:05:40.192 real 0m6.962s 00:05:40.192 user 0m6.496s 00:05:40.192 sys 0m0.733s 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:40.192 ************************************ 00:05:40.192 END TEST skip_rpc_with_json 00:05:40.192 ************************************ 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:40.192 16:44:01 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:05:40.192 16:44:01 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:40.192 16:44:01 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:40.192 16:44:01 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:40.192 ************************************ 00:05:40.192 START TEST skip_rpc_with_delay 00:05:40.192 ************************************ 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_delay 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # local es=0 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:40.192 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:05:40.451 [2024-09-29 16:44:01.865374] app.c: 840:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:05:40.451 [2024-09-29 16:44:01.865501] app.c: 719:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 0, errno: 2 00:05:40.451 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # es=1 00:05:40.451 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:40.451 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:05:40.451 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:40.451 ************************************ 00:05:40.451 END TEST skip_rpc_with_delay 00:05:40.451 ************************************ 00:05:40.451 00:05:40.451 real 0m0.157s 00:05:40.451 user 0m0.088s 00:05:40.451 sys 0m0.068s 00:05:40.451 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:40.451 16:44:01 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:05:40.451 16:44:01 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:05:40.451 16:44:01 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:05:40.451 16:44:01 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:05:40.451 16:44:01 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:40.451 16:44:01 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:40.451 16:44:01 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:40.451 ************************************ 00:05:40.451 START TEST exit_on_failed_rpc_init 00:05:40.452 ************************************ 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1125 -- # test_exit_on_failed_rpc_init 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69241 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69241 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@831 -- # '[' -z 69241 ']' 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:40.452 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:40.452 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:40.452 [2024-09-29 16:44:02.090007] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:40.452 [2024-09-29 16:44:02.090240] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69241 ] 00:05:40.711 [2024-09-29 16:44:02.232525] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:40.711 [2024-09-29 16:44:02.277938] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # return 0 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # local es=0 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:05:41.280 16:44:02 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:05:41.539 [2024-09-29 16:44:03.002799] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:41.539 [2024-09-29 16:44:03.003007] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69259 ] 00:05:41.539 [2024-09-29 16:44:03.149892] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:41.539 [2024-09-29 16:44:03.197632] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:41.539 [2024-09-29 16:44:03.197828] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:05:41.539 [2024-09-29 16:44:03.197881] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:05:41.539 [2024-09-29 16:44:03.197904] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # es=234 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@662 -- # es=106 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # case "$es" in 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@670 -- # es=1 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69241 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@950 -- # '[' -z 69241 ']' 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # kill -0 69241 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # uname 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69241 00:05:41.799 killing process with pid 69241 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69241' 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@969 -- # kill 69241 00:05:41.799 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@974 -- # wait 69241 00:05:42.368 ************************************ 00:05:42.368 END TEST exit_on_failed_rpc_init 00:05:42.368 ************************************ 00:05:42.368 00:05:42.368 real 0m1.739s 00:05:42.368 user 0m1.871s 00:05:42.368 sys 0m0.491s 00:05:42.368 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:42.368 16:44:03 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:05:42.368 16:44:03 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:42.368 00:05:42.368 real 0m14.809s 00:05:42.368 user 0m13.710s 00:05:42.368 sys 0m1.919s 00:05:42.368 ************************************ 00:05:42.368 END TEST skip_rpc 00:05:42.368 ************************************ 00:05:42.368 16:44:03 skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:42.368 16:44:03 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:42.368 16:44:03 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:42.368 16:44:03 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:42.368 16:44:03 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:42.368 16:44:03 -- common/autotest_common.sh@10 -- # set +x 00:05:42.368 ************************************ 00:05:42.368 START TEST rpc_client 00:05:42.368 ************************************ 00:05:42.368 16:44:03 rpc_client -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:05:42.368 * Looking for test storage... 00:05:42.368 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:05:42.368 16:44:03 rpc_client -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:42.368 16:44:03 rpc_client -- common/autotest_common.sh@1681 -- # lcov --version 00:05:42.368 16:44:03 rpc_client -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@345 -- # : 1 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@353 -- # local d=1 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@355 -- # echo 1 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@353 -- # local d=2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@355 -- # echo 2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:42.628 16:44:04 rpc_client -- scripts/common.sh@368 -- # return 0 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:42.628 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.628 --rc genhtml_branch_coverage=1 00:05:42.628 --rc genhtml_function_coverage=1 00:05:42.628 --rc genhtml_legend=1 00:05:42.628 --rc geninfo_all_blocks=1 00:05:42.628 --rc geninfo_unexecuted_blocks=1 00:05:42.628 00:05:42.628 ' 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:42.628 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.628 --rc genhtml_branch_coverage=1 00:05:42.628 --rc genhtml_function_coverage=1 00:05:42.628 --rc genhtml_legend=1 00:05:42.628 --rc geninfo_all_blocks=1 00:05:42.628 --rc geninfo_unexecuted_blocks=1 00:05:42.628 00:05:42.628 ' 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:42.628 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.628 --rc genhtml_branch_coverage=1 00:05:42.628 --rc genhtml_function_coverage=1 00:05:42.628 --rc genhtml_legend=1 00:05:42.628 --rc geninfo_all_blocks=1 00:05:42.628 --rc geninfo_unexecuted_blocks=1 00:05:42.628 00:05:42.628 ' 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:42.628 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.628 --rc genhtml_branch_coverage=1 00:05:42.628 --rc genhtml_function_coverage=1 00:05:42.628 --rc genhtml_legend=1 00:05:42.628 --rc geninfo_all_blocks=1 00:05:42.628 --rc geninfo_unexecuted_blocks=1 00:05:42.628 00:05:42.628 ' 00:05:42.628 16:44:04 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:05:42.628 OK 00:05:42.628 16:44:04 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:05:42.628 00:05:42.628 real 0m0.280s 00:05:42.628 user 0m0.143s 00:05:42.628 sys 0m0.153s 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:42.628 16:44:04 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:05:42.628 ************************************ 00:05:42.628 END TEST rpc_client 00:05:42.628 ************************************ 00:05:42.628 16:44:04 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:42.628 16:44:04 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:42.628 16:44:04 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:42.628 16:44:04 -- common/autotest_common.sh@10 -- # set +x 00:05:42.629 ************************************ 00:05:42.629 START TEST json_config 00:05:42.629 ************************************ 00:05:42.629 16:44:04 json_config -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:05:42.629 16:44:04 json_config -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1681 -- # lcov --version 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:42.889 16:44:04 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:42.889 16:44:04 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:05:42.889 16:44:04 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:05:42.889 16:44:04 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:05:42.889 16:44:04 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:05:42.889 16:44:04 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:05:42.889 16:44:04 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:42.889 16:44:04 json_config -- scripts/common.sh@344 -- # case "$op" in 00:05:42.889 16:44:04 json_config -- scripts/common.sh@345 -- # : 1 00:05:42.889 16:44:04 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:42.889 16:44:04 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:42.889 16:44:04 json_config -- scripts/common.sh@365 -- # decimal 1 00:05:42.889 16:44:04 json_config -- scripts/common.sh@353 -- # local d=1 00:05:42.889 16:44:04 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:42.889 16:44:04 json_config -- scripts/common.sh@355 -- # echo 1 00:05:42.889 16:44:04 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:05:42.889 16:44:04 json_config -- scripts/common.sh@366 -- # decimal 2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@353 -- # local d=2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:42.889 16:44:04 json_config -- scripts/common.sh@355 -- # echo 2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:05:42.889 16:44:04 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:42.889 16:44:04 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:42.889 16:44:04 json_config -- scripts/common.sh@368 -- # return 0 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:42.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.889 --rc genhtml_branch_coverage=1 00:05:42.889 --rc genhtml_function_coverage=1 00:05:42.889 --rc genhtml_legend=1 00:05:42.889 --rc geninfo_all_blocks=1 00:05:42.889 --rc geninfo_unexecuted_blocks=1 00:05:42.889 00:05:42.889 ' 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:42.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.889 --rc genhtml_branch_coverage=1 00:05:42.889 --rc genhtml_function_coverage=1 00:05:42.889 --rc genhtml_legend=1 00:05:42.889 --rc geninfo_all_blocks=1 00:05:42.889 --rc geninfo_unexecuted_blocks=1 00:05:42.889 00:05:42.889 ' 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:42.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.889 --rc genhtml_branch_coverage=1 00:05:42.889 --rc genhtml_function_coverage=1 00:05:42.889 --rc genhtml_legend=1 00:05:42.889 --rc geninfo_all_blocks=1 00:05:42.889 --rc geninfo_unexecuted_blocks=1 00:05:42.889 00:05:42.889 ' 00:05:42.889 16:44:04 json_config -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:42.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.889 --rc genhtml_branch_coverage=1 00:05:42.889 --rc genhtml_function_coverage=1 00:05:42.889 --rc genhtml_legend=1 00:05:42.889 --rc geninfo_all_blocks=1 00:05:42.889 --rc geninfo_unexecuted_blocks=1 00:05:42.889 00:05:42.889 ' 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@7 -- # uname -s 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:015174e6-42cc-4d4f-853e-c7d76863366c 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=015174e6-42cc-4d4f-853e-c7d76863366c 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:42.890 16:44:04 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:05:42.890 16:44:04 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:42.890 16:44:04 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:42.890 16:44:04 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:42.890 16:44:04 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.890 16:44:04 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.890 16:44:04 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.890 16:44:04 json_config -- paths/export.sh@5 -- # export PATH 00:05:42.890 16:44:04 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@51 -- # : 0 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:42.890 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:42.890 16:44:04 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:05:42.890 WARNING: No tests are enabled so not running JSON configuration tests 00:05:42.890 16:44:04 json_config -- json_config/json_config.sh@28 -- # exit 0 00:05:42.890 00:05:42.890 real 0m0.236s 00:05:42.890 user 0m0.145s 00:05:42.890 sys 0m0.094s 00:05:42.890 16:44:04 json_config -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:42.890 16:44:04 json_config -- common/autotest_common.sh@10 -- # set +x 00:05:42.890 ************************************ 00:05:42.890 END TEST json_config 00:05:42.890 ************************************ 00:05:42.890 16:44:04 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:42.890 16:44:04 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:42.890 16:44:04 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:42.890 16:44:04 -- common/autotest_common.sh@10 -- # set +x 00:05:42.890 ************************************ 00:05:42.890 START TEST json_config_extra_key 00:05:42.890 ************************************ 00:05:42.890 16:44:04 json_config_extra_key -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:05:43.151 16:44:04 json_config_extra_key -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:43.151 16:44:04 json_config_extra_key -- common/autotest_common.sh@1681 -- # lcov --version 00:05:43.151 16:44:04 json_config_extra_key -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:43.151 16:44:04 json_config_extra_key -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:05:43.151 16:44:04 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:43.152 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.152 --rc genhtml_branch_coverage=1 00:05:43.152 --rc genhtml_function_coverage=1 00:05:43.152 --rc genhtml_legend=1 00:05:43.152 --rc geninfo_all_blocks=1 00:05:43.152 --rc geninfo_unexecuted_blocks=1 00:05:43.152 00:05:43.152 ' 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:43.152 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.152 --rc genhtml_branch_coverage=1 00:05:43.152 --rc genhtml_function_coverage=1 00:05:43.152 --rc genhtml_legend=1 00:05:43.152 --rc geninfo_all_blocks=1 00:05:43.152 --rc geninfo_unexecuted_blocks=1 00:05:43.152 00:05:43.152 ' 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:43.152 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.152 --rc genhtml_branch_coverage=1 00:05:43.152 --rc genhtml_function_coverage=1 00:05:43.152 --rc genhtml_legend=1 00:05:43.152 --rc geninfo_all_blocks=1 00:05:43.152 --rc geninfo_unexecuted_blocks=1 00:05:43.152 00:05:43.152 ' 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:43.152 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:43.152 --rc genhtml_branch_coverage=1 00:05:43.152 --rc genhtml_function_coverage=1 00:05:43.152 --rc genhtml_legend=1 00:05:43.152 --rc geninfo_all_blocks=1 00:05:43.152 --rc geninfo_unexecuted_blocks=1 00:05:43.152 00:05:43.152 ' 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:015174e6-42cc-4d4f-853e-c7d76863366c 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=015174e6-42cc-4d4f-853e-c7d76863366c 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:43.152 16:44:04 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:43.152 16:44:04 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:43.152 16:44:04 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:43.152 16:44:04 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:43.152 16:44:04 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:05:43.152 16:44:04 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:43.152 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:43.152 16:44:04 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:05:43.152 INFO: launching applications... 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:05:43.152 16:44:04 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69441 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:05:43.152 Waiting for target to run... 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69441 /var/tmp/spdk_tgt.sock 00:05:43.152 16:44:04 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@831 -- # '[' -z 69441 ']' 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:43.152 16:44:04 json_config_extra_key -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:05:43.152 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:05:43.153 16:44:04 json_config_extra_key -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:43.153 16:44:04 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:43.413 [2024-09-29 16:44:04.834288] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:43.413 [2024-09-29 16:44:04.834491] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69441 ] 00:05:43.673 [2024-09-29 16:44:05.205905] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:43.673 [2024-09-29 16:44:05.235703] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:44.242 00:05:44.242 INFO: shutting down applications... 00:05:44.242 16:44:05 json_config_extra_key -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:44.242 16:44:05 json_config_extra_key -- common/autotest_common.sh@864 -- # return 0 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:05:44.242 16:44:05 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:05:44.242 16:44:05 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69441 ]] 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69441 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69441 00:05:44.242 16:44:05 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:05:44.501 16:44:06 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:05:44.501 16:44:06 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:05:44.501 16:44:06 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69441 00:05:44.501 16:44:06 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:05:44.501 16:44:06 json_config_extra_key -- json_config/common.sh@43 -- # break 00:05:44.501 16:44:06 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:05:44.501 16:44:06 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:05:44.501 SPDK target shutdown done 00:05:44.501 16:44:06 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:05:44.501 Success 00:05:44.501 00:05:44.501 real 0m1.637s 00:05:44.501 user 0m1.323s 00:05:44.501 sys 0m0.484s 00:05:44.501 16:44:06 json_config_extra_key -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:44.501 16:44:06 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:05:44.501 ************************************ 00:05:44.501 END TEST json_config_extra_key 00:05:44.501 ************************************ 00:05:44.760 16:44:06 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:44.760 16:44:06 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:44.760 16:44:06 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:44.760 16:44:06 -- common/autotest_common.sh@10 -- # set +x 00:05:44.760 ************************************ 00:05:44.760 START TEST alias_rpc 00:05:44.760 ************************************ 00:05:44.760 16:44:06 alias_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:05:44.760 * Looking for test storage... 00:05:44.760 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:05:44.760 16:44:06 alias_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:44.760 16:44:06 alias_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:44.760 16:44:06 alias_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:44.760 16:44:06 alias_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@345 -- # : 1 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:44.760 16:44:06 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:45.020 16:44:06 alias_rpc -- scripts/common.sh@368 -- # return 0 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:45.020 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.020 --rc genhtml_branch_coverage=1 00:05:45.020 --rc genhtml_function_coverage=1 00:05:45.020 --rc genhtml_legend=1 00:05:45.020 --rc geninfo_all_blocks=1 00:05:45.020 --rc geninfo_unexecuted_blocks=1 00:05:45.020 00:05:45.020 ' 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:45.020 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.020 --rc genhtml_branch_coverage=1 00:05:45.020 --rc genhtml_function_coverage=1 00:05:45.020 --rc genhtml_legend=1 00:05:45.020 --rc geninfo_all_blocks=1 00:05:45.020 --rc geninfo_unexecuted_blocks=1 00:05:45.020 00:05:45.020 ' 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:45.020 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.020 --rc genhtml_branch_coverage=1 00:05:45.020 --rc genhtml_function_coverage=1 00:05:45.020 --rc genhtml_legend=1 00:05:45.020 --rc geninfo_all_blocks=1 00:05:45.020 --rc geninfo_unexecuted_blocks=1 00:05:45.020 00:05:45.020 ' 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:45.020 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:45.020 --rc genhtml_branch_coverage=1 00:05:45.020 --rc genhtml_function_coverage=1 00:05:45.020 --rc genhtml_legend=1 00:05:45.020 --rc geninfo_all_blocks=1 00:05:45.020 --rc geninfo_unexecuted_blocks=1 00:05:45.020 00:05:45.020 ' 00:05:45.020 16:44:06 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:05:45.020 16:44:06 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:45.020 16:44:06 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=69515 00:05:45.020 16:44:06 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 69515 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@831 -- # '[' -z 69515 ']' 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:45.020 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:45.020 16:44:06 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:45.020 [2024-09-29 16:44:06.517241] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:45.020 [2024-09-29 16:44:06.517801] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69515 ] 00:05:45.020 [2024-09-29 16:44:06.661293] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:45.280 [2024-09-29 16:44:06.707273] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:45.848 16:44:07 alias_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:45.848 16:44:07 alias_rpc -- common/autotest_common.sh@864 -- # return 0 00:05:45.848 16:44:07 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:05:46.108 16:44:07 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 69515 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@950 -- # '[' -z 69515 ']' 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@954 -- # kill -0 69515 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@955 -- # uname 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69515 00:05:46.108 killing process with pid 69515 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69515' 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@969 -- # kill 69515 00:05:46.108 16:44:07 alias_rpc -- common/autotest_common.sh@974 -- # wait 69515 00:05:46.368 ************************************ 00:05:46.368 END TEST alias_rpc 00:05:46.368 ************************************ 00:05:46.368 00:05:46.368 real 0m1.728s 00:05:46.368 user 0m1.712s 00:05:46.368 sys 0m0.491s 00:05:46.368 16:44:07 alias_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:46.368 16:44:07 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:46.368 16:44:07 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:05:46.368 16:44:07 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:46.368 16:44:07 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:46.368 16:44:07 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:46.368 16:44:07 -- common/autotest_common.sh@10 -- # set +x 00:05:46.368 ************************************ 00:05:46.368 START TEST spdkcli_tcp 00:05:46.368 ************************************ 00:05:46.368 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:05:46.629 * Looking for test storage... 00:05:46.629 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:05:46.629 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:46.629 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lcov --version 00:05:46.629 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:46.629 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:46.629 16:44:08 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:46.629 16:44:08 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:46.629 16:44:08 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:46.630 16:44:08 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:46.630 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.630 --rc genhtml_branch_coverage=1 00:05:46.630 --rc genhtml_function_coverage=1 00:05:46.630 --rc genhtml_legend=1 00:05:46.630 --rc geninfo_all_blocks=1 00:05:46.630 --rc geninfo_unexecuted_blocks=1 00:05:46.630 00:05:46.630 ' 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:46.630 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.630 --rc genhtml_branch_coverage=1 00:05:46.630 --rc genhtml_function_coverage=1 00:05:46.630 --rc genhtml_legend=1 00:05:46.630 --rc geninfo_all_blocks=1 00:05:46.630 --rc geninfo_unexecuted_blocks=1 00:05:46.630 00:05:46.630 ' 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:46.630 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.630 --rc genhtml_branch_coverage=1 00:05:46.630 --rc genhtml_function_coverage=1 00:05:46.630 --rc genhtml_legend=1 00:05:46.630 --rc geninfo_all_blocks=1 00:05:46.630 --rc geninfo_unexecuted_blocks=1 00:05:46.630 00:05:46.630 ' 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:46.630 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.630 --rc genhtml_branch_coverage=1 00:05:46.630 --rc genhtml_function_coverage=1 00:05:46.630 --rc genhtml_legend=1 00:05:46.630 --rc geninfo_all_blocks=1 00:05:46.630 --rc geninfo_unexecuted_blocks=1 00:05:46.630 00:05:46.630 ' 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=69600 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:05:46.630 16:44:08 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 69600 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@831 -- # '[' -z 69600 ']' 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:46.630 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:46.630 16:44:08 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:46.930 [2024-09-29 16:44:08.332369] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:46.930 [2024-09-29 16:44:08.332584] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69600 ] 00:05:46.930 [2024-09-29 16:44:08.478641] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:46.930 [2024-09-29 16:44:08.524469] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:46.930 [2024-09-29 16:44:08.524560] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:47.575 16:44:09 spdkcli_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:47.575 16:44:09 spdkcli_tcp -- common/autotest_common.sh@864 -- # return 0 00:05:47.575 16:44:09 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=69617 00:05:47.575 16:44:09 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:05:47.575 16:44:09 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:05:47.835 [ 00:05:47.835 "bdev_malloc_delete", 00:05:47.835 "bdev_malloc_create", 00:05:47.835 "bdev_null_resize", 00:05:47.835 "bdev_null_delete", 00:05:47.835 "bdev_null_create", 00:05:47.835 "bdev_nvme_cuse_unregister", 00:05:47.835 "bdev_nvme_cuse_register", 00:05:47.835 "bdev_opal_new_user", 00:05:47.835 "bdev_opal_set_lock_state", 00:05:47.835 "bdev_opal_delete", 00:05:47.835 "bdev_opal_get_info", 00:05:47.835 "bdev_opal_create", 00:05:47.835 "bdev_nvme_opal_revert", 00:05:47.835 "bdev_nvme_opal_init", 00:05:47.835 "bdev_nvme_send_cmd", 00:05:47.835 "bdev_nvme_set_keys", 00:05:47.835 "bdev_nvme_get_path_iostat", 00:05:47.835 "bdev_nvme_get_mdns_discovery_info", 00:05:47.835 "bdev_nvme_stop_mdns_discovery", 00:05:47.835 "bdev_nvme_start_mdns_discovery", 00:05:47.835 "bdev_nvme_set_multipath_policy", 00:05:47.835 "bdev_nvme_set_preferred_path", 00:05:47.835 "bdev_nvme_get_io_paths", 00:05:47.835 "bdev_nvme_remove_error_injection", 00:05:47.835 "bdev_nvme_add_error_injection", 00:05:47.835 "bdev_nvme_get_discovery_info", 00:05:47.835 "bdev_nvme_stop_discovery", 00:05:47.835 "bdev_nvme_start_discovery", 00:05:47.835 "bdev_nvme_get_controller_health_info", 00:05:47.835 "bdev_nvme_disable_controller", 00:05:47.835 "bdev_nvme_enable_controller", 00:05:47.835 "bdev_nvme_reset_controller", 00:05:47.835 "bdev_nvme_get_transport_statistics", 00:05:47.835 "bdev_nvme_apply_firmware", 00:05:47.836 "bdev_nvme_detach_controller", 00:05:47.836 "bdev_nvme_get_controllers", 00:05:47.836 "bdev_nvme_attach_controller", 00:05:47.836 "bdev_nvme_set_hotplug", 00:05:47.836 "bdev_nvme_set_options", 00:05:47.836 "bdev_passthru_delete", 00:05:47.836 "bdev_passthru_create", 00:05:47.836 "bdev_lvol_set_parent_bdev", 00:05:47.836 "bdev_lvol_set_parent", 00:05:47.836 "bdev_lvol_check_shallow_copy", 00:05:47.836 "bdev_lvol_start_shallow_copy", 00:05:47.836 "bdev_lvol_grow_lvstore", 00:05:47.836 "bdev_lvol_get_lvols", 00:05:47.836 "bdev_lvol_get_lvstores", 00:05:47.836 "bdev_lvol_delete", 00:05:47.836 "bdev_lvol_set_read_only", 00:05:47.836 "bdev_lvol_resize", 00:05:47.836 "bdev_lvol_decouple_parent", 00:05:47.836 "bdev_lvol_inflate", 00:05:47.836 "bdev_lvol_rename", 00:05:47.836 "bdev_lvol_clone_bdev", 00:05:47.836 "bdev_lvol_clone", 00:05:47.836 "bdev_lvol_snapshot", 00:05:47.836 "bdev_lvol_create", 00:05:47.836 "bdev_lvol_delete_lvstore", 00:05:47.836 "bdev_lvol_rename_lvstore", 00:05:47.836 "bdev_lvol_create_lvstore", 00:05:47.836 "bdev_raid_set_options", 00:05:47.836 "bdev_raid_remove_base_bdev", 00:05:47.836 "bdev_raid_add_base_bdev", 00:05:47.836 "bdev_raid_delete", 00:05:47.836 "bdev_raid_create", 00:05:47.836 "bdev_raid_get_bdevs", 00:05:47.836 "bdev_error_inject_error", 00:05:47.836 "bdev_error_delete", 00:05:47.836 "bdev_error_create", 00:05:47.836 "bdev_split_delete", 00:05:47.836 "bdev_split_create", 00:05:47.836 "bdev_delay_delete", 00:05:47.836 "bdev_delay_create", 00:05:47.836 "bdev_delay_update_latency", 00:05:47.836 "bdev_zone_block_delete", 00:05:47.836 "bdev_zone_block_create", 00:05:47.836 "blobfs_create", 00:05:47.836 "blobfs_detect", 00:05:47.836 "blobfs_set_cache_size", 00:05:47.836 "bdev_aio_delete", 00:05:47.836 "bdev_aio_rescan", 00:05:47.836 "bdev_aio_create", 00:05:47.836 "bdev_ftl_set_property", 00:05:47.836 "bdev_ftl_get_properties", 00:05:47.836 "bdev_ftl_get_stats", 00:05:47.836 "bdev_ftl_unmap", 00:05:47.836 "bdev_ftl_unload", 00:05:47.836 "bdev_ftl_delete", 00:05:47.836 "bdev_ftl_load", 00:05:47.836 "bdev_ftl_create", 00:05:47.836 "bdev_virtio_attach_controller", 00:05:47.836 "bdev_virtio_scsi_get_devices", 00:05:47.836 "bdev_virtio_detach_controller", 00:05:47.836 "bdev_virtio_blk_set_hotplug", 00:05:47.836 "bdev_iscsi_delete", 00:05:47.836 "bdev_iscsi_create", 00:05:47.836 "bdev_iscsi_set_options", 00:05:47.836 "accel_error_inject_error", 00:05:47.836 "ioat_scan_accel_module", 00:05:47.836 "dsa_scan_accel_module", 00:05:47.836 "iaa_scan_accel_module", 00:05:47.836 "keyring_file_remove_key", 00:05:47.836 "keyring_file_add_key", 00:05:47.836 "keyring_linux_set_options", 00:05:47.836 "fsdev_aio_delete", 00:05:47.836 "fsdev_aio_create", 00:05:47.836 "iscsi_get_histogram", 00:05:47.836 "iscsi_enable_histogram", 00:05:47.836 "iscsi_set_options", 00:05:47.836 "iscsi_get_auth_groups", 00:05:47.836 "iscsi_auth_group_remove_secret", 00:05:47.836 "iscsi_auth_group_add_secret", 00:05:47.836 "iscsi_delete_auth_group", 00:05:47.836 "iscsi_create_auth_group", 00:05:47.836 "iscsi_set_discovery_auth", 00:05:47.836 "iscsi_get_options", 00:05:47.836 "iscsi_target_node_request_logout", 00:05:47.836 "iscsi_target_node_set_redirect", 00:05:47.836 "iscsi_target_node_set_auth", 00:05:47.836 "iscsi_target_node_add_lun", 00:05:47.836 "iscsi_get_stats", 00:05:47.836 "iscsi_get_connections", 00:05:47.836 "iscsi_portal_group_set_auth", 00:05:47.836 "iscsi_start_portal_group", 00:05:47.836 "iscsi_delete_portal_group", 00:05:47.836 "iscsi_create_portal_group", 00:05:47.836 "iscsi_get_portal_groups", 00:05:47.836 "iscsi_delete_target_node", 00:05:47.836 "iscsi_target_node_remove_pg_ig_maps", 00:05:47.836 "iscsi_target_node_add_pg_ig_maps", 00:05:47.836 "iscsi_create_target_node", 00:05:47.836 "iscsi_get_target_nodes", 00:05:47.836 "iscsi_delete_initiator_group", 00:05:47.836 "iscsi_initiator_group_remove_initiators", 00:05:47.836 "iscsi_initiator_group_add_initiators", 00:05:47.836 "iscsi_create_initiator_group", 00:05:47.836 "iscsi_get_initiator_groups", 00:05:47.836 "nvmf_set_crdt", 00:05:47.836 "nvmf_set_config", 00:05:47.836 "nvmf_set_max_subsystems", 00:05:47.836 "nvmf_stop_mdns_prr", 00:05:47.836 "nvmf_publish_mdns_prr", 00:05:47.836 "nvmf_subsystem_get_listeners", 00:05:47.836 "nvmf_subsystem_get_qpairs", 00:05:47.836 "nvmf_subsystem_get_controllers", 00:05:47.836 "nvmf_get_stats", 00:05:47.836 "nvmf_get_transports", 00:05:47.836 "nvmf_create_transport", 00:05:47.836 "nvmf_get_targets", 00:05:47.836 "nvmf_delete_target", 00:05:47.836 "nvmf_create_target", 00:05:47.836 "nvmf_subsystem_allow_any_host", 00:05:47.836 "nvmf_subsystem_set_keys", 00:05:47.836 "nvmf_subsystem_remove_host", 00:05:47.836 "nvmf_subsystem_add_host", 00:05:47.836 "nvmf_ns_remove_host", 00:05:47.836 "nvmf_ns_add_host", 00:05:47.836 "nvmf_subsystem_remove_ns", 00:05:47.836 "nvmf_subsystem_set_ns_ana_group", 00:05:47.836 "nvmf_subsystem_add_ns", 00:05:47.836 "nvmf_subsystem_listener_set_ana_state", 00:05:47.836 "nvmf_discovery_get_referrals", 00:05:47.836 "nvmf_discovery_remove_referral", 00:05:47.836 "nvmf_discovery_add_referral", 00:05:47.836 "nvmf_subsystem_remove_listener", 00:05:47.836 "nvmf_subsystem_add_listener", 00:05:47.836 "nvmf_delete_subsystem", 00:05:47.836 "nvmf_create_subsystem", 00:05:47.836 "nvmf_get_subsystems", 00:05:47.836 "env_dpdk_get_mem_stats", 00:05:47.836 "nbd_get_disks", 00:05:47.836 "nbd_stop_disk", 00:05:47.836 "nbd_start_disk", 00:05:47.836 "ublk_recover_disk", 00:05:47.836 "ublk_get_disks", 00:05:47.836 "ublk_stop_disk", 00:05:47.836 "ublk_start_disk", 00:05:47.836 "ublk_destroy_target", 00:05:47.836 "ublk_create_target", 00:05:47.836 "virtio_blk_create_transport", 00:05:47.836 "virtio_blk_get_transports", 00:05:47.836 "vhost_controller_set_coalescing", 00:05:47.836 "vhost_get_controllers", 00:05:47.836 "vhost_delete_controller", 00:05:47.836 "vhost_create_blk_controller", 00:05:47.836 "vhost_scsi_controller_remove_target", 00:05:47.836 "vhost_scsi_controller_add_target", 00:05:47.836 "vhost_start_scsi_controller", 00:05:47.836 "vhost_create_scsi_controller", 00:05:47.836 "thread_set_cpumask", 00:05:47.836 "scheduler_set_options", 00:05:47.836 "framework_get_governor", 00:05:47.836 "framework_get_scheduler", 00:05:47.836 "framework_set_scheduler", 00:05:47.836 "framework_get_reactors", 00:05:47.836 "thread_get_io_channels", 00:05:47.836 "thread_get_pollers", 00:05:47.836 "thread_get_stats", 00:05:47.836 "framework_monitor_context_switch", 00:05:47.836 "spdk_kill_instance", 00:05:47.836 "log_enable_timestamps", 00:05:47.836 "log_get_flags", 00:05:47.836 "log_clear_flag", 00:05:47.836 "log_set_flag", 00:05:47.836 "log_get_level", 00:05:47.836 "log_set_level", 00:05:47.836 "log_get_print_level", 00:05:47.836 "log_set_print_level", 00:05:47.836 "framework_enable_cpumask_locks", 00:05:47.836 "framework_disable_cpumask_locks", 00:05:47.836 "framework_wait_init", 00:05:47.836 "framework_start_init", 00:05:47.836 "scsi_get_devices", 00:05:47.836 "bdev_get_histogram", 00:05:47.836 "bdev_enable_histogram", 00:05:47.836 "bdev_set_qos_limit", 00:05:47.836 "bdev_set_qd_sampling_period", 00:05:47.836 "bdev_get_bdevs", 00:05:47.836 "bdev_reset_iostat", 00:05:47.836 "bdev_get_iostat", 00:05:47.836 "bdev_examine", 00:05:47.836 "bdev_wait_for_examine", 00:05:47.836 "bdev_set_options", 00:05:47.836 "accel_get_stats", 00:05:47.836 "accel_set_options", 00:05:47.836 "accel_set_driver", 00:05:47.836 "accel_crypto_key_destroy", 00:05:47.836 "accel_crypto_keys_get", 00:05:47.836 "accel_crypto_key_create", 00:05:47.836 "accel_assign_opc", 00:05:47.836 "accel_get_module_info", 00:05:47.836 "accel_get_opc_assignments", 00:05:47.836 "vmd_rescan", 00:05:47.836 "vmd_remove_device", 00:05:47.836 "vmd_enable", 00:05:47.836 "sock_get_default_impl", 00:05:47.836 "sock_set_default_impl", 00:05:47.836 "sock_impl_set_options", 00:05:47.836 "sock_impl_get_options", 00:05:47.836 "iobuf_get_stats", 00:05:47.836 "iobuf_set_options", 00:05:47.836 "keyring_get_keys", 00:05:47.836 "framework_get_pci_devices", 00:05:47.836 "framework_get_config", 00:05:47.836 "framework_get_subsystems", 00:05:47.836 "fsdev_set_opts", 00:05:47.836 "fsdev_get_opts", 00:05:47.836 "trace_get_info", 00:05:47.836 "trace_get_tpoint_group_mask", 00:05:47.836 "trace_disable_tpoint_group", 00:05:47.836 "trace_enable_tpoint_group", 00:05:47.836 "trace_clear_tpoint_mask", 00:05:47.836 "trace_set_tpoint_mask", 00:05:47.836 "notify_get_notifications", 00:05:47.836 "notify_get_types", 00:05:47.836 "spdk_get_version", 00:05:47.836 "rpc_get_methods" 00:05:47.836 ] 00:05:47.836 16:44:09 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:05:47.836 16:44:09 spdkcli_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:47.836 16:44:09 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:47.836 16:44:09 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:05:47.836 16:44:09 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 69600 00:05:47.836 16:44:09 spdkcli_tcp -- common/autotest_common.sh@950 -- # '[' -z 69600 ']' 00:05:47.836 16:44:09 spdkcli_tcp -- common/autotest_common.sh@954 -- # kill -0 69600 00:05:47.836 16:44:09 spdkcli_tcp -- common/autotest_common.sh@955 -- # uname 00:05:47.836 16:44:09 spdkcli_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:47.836 16:44:09 spdkcli_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69600 00:05:47.837 killing process with pid 69600 00:05:47.837 16:44:09 spdkcli_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:47.837 16:44:09 spdkcli_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:47.837 16:44:09 spdkcli_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69600' 00:05:47.837 16:44:09 spdkcli_tcp -- common/autotest_common.sh@969 -- # kill 69600 00:05:47.837 16:44:09 spdkcli_tcp -- common/autotest_common.sh@974 -- # wait 69600 00:05:48.775 ************************************ 00:05:48.775 END TEST spdkcli_tcp 00:05:48.775 ************************************ 00:05:48.775 00:05:48.775 real 0m2.096s 00:05:48.775 user 0m3.492s 00:05:48.775 sys 0m0.573s 00:05:48.775 16:44:10 spdkcli_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:48.775 16:44:10 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:05:48.775 16:44:10 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:48.775 16:44:10 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:48.775 16:44:10 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:48.775 16:44:10 -- common/autotest_common.sh@10 -- # set +x 00:05:48.775 ************************************ 00:05:48.775 START TEST dpdk_mem_utility 00:05:48.775 ************************************ 00:05:48.775 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:05:48.775 * Looking for test storage... 00:05:48.775 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:05:48.775 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:48.775 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lcov --version 00:05:48.775 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:48.775 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:05:48.775 16:44:10 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:48.776 16:44:10 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:48.776 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.776 --rc genhtml_branch_coverage=1 00:05:48.776 --rc genhtml_function_coverage=1 00:05:48.776 --rc genhtml_legend=1 00:05:48.776 --rc geninfo_all_blocks=1 00:05:48.776 --rc geninfo_unexecuted_blocks=1 00:05:48.776 00:05:48.776 ' 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:48.776 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.776 --rc genhtml_branch_coverage=1 00:05:48.776 --rc genhtml_function_coverage=1 00:05:48.776 --rc genhtml_legend=1 00:05:48.776 --rc geninfo_all_blocks=1 00:05:48.776 --rc geninfo_unexecuted_blocks=1 00:05:48.776 00:05:48.776 ' 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:48.776 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.776 --rc genhtml_branch_coverage=1 00:05:48.776 --rc genhtml_function_coverage=1 00:05:48.776 --rc genhtml_legend=1 00:05:48.776 --rc geninfo_all_blocks=1 00:05:48.776 --rc geninfo_unexecuted_blocks=1 00:05:48.776 00:05:48.776 ' 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:48.776 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:48.776 --rc genhtml_branch_coverage=1 00:05:48.776 --rc genhtml_function_coverage=1 00:05:48.776 --rc genhtml_legend=1 00:05:48.776 --rc geninfo_all_blocks=1 00:05:48.776 --rc geninfo_unexecuted_blocks=1 00:05:48.776 00:05:48.776 ' 00:05:48.776 16:44:10 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:48.776 16:44:10 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=69700 00:05:48.776 16:44:10 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:05:48.776 16:44:10 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 69700 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@831 -- # '[' -z 69700 ']' 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:48.776 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:48.776 16:44:10 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:49.036 [2024-09-29 16:44:10.481320] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:49.036 [2024-09-29 16:44:10.481460] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69700 ] 00:05:49.036 [2024-09-29 16:44:10.628555] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:49.036 [2024-09-29 16:44:10.704575] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:49.978 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:49.978 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@864 -- # return 0 00:05:49.978 16:44:11 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:05:49.978 16:44:11 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:05:49.978 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:49.978 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:49.978 { 00:05:49.978 "filename": "/tmp/spdk_mem_dump.txt" 00:05:49.978 } 00:05:49.978 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:49.978 16:44:11 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:05:49.978 DPDK memory size 860.000000 MiB in 1 heap(s) 00:05:49.978 1 heaps totaling size 860.000000 MiB 00:05:49.978 size: 860.000000 MiB heap id: 0 00:05:49.978 end heaps---------- 00:05:49.978 9 mempools totaling size 642.649841 MiB 00:05:49.978 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:05:49.978 size: 158.602051 MiB name: PDU_data_out_Pool 00:05:49.978 size: 92.545471 MiB name: bdev_io_69700 00:05:49.978 size: 51.011292 MiB name: evtpool_69700 00:05:49.978 size: 50.003479 MiB name: msgpool_69700 00:05:49.978 size: 36.509338 MiB name: fsdev_io_69700 00:05:49.978 size: 21.763794 MiB name: PDU_Pool 00:05:49.978 size: 19.513306 MiB name: SCSI_TASK_Pool 00:05:49.978 size: 0.026123 MiB name: Session_Pool 00:05:49.978 end mempools------- 00:05:49.978 6 memzones totaling size 4.142822 MiB 00:05:49.978 size: 1.000366 MiB name: RG_ring_0_69700 00:05:49.978 size: 1.000366 MiB name: RG_ring_1_69700 00:05:49.978 size: 1.000366 MiB name: RG_ring_4_69700 00:05:49.978 size: 1.000366 MiB name: RG_ring_5_69700 00:05:49.978 size: 0.125366 MiB name: RG_ring_2_69700 00:05:49.978 size: 0.015991 MiB name: RG_ring_3_69700 00:05:49.978 end memzones------- 00:05:49.978 16:44:11 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:05:49.978 heap id: 0 total size: 860.000000 MiB number of busy elements: 303 number of free elements: 16 00:05:49.978 list of free elements. size: 13.937256 MiB 00:05:49.978 element at address: 0x200000400000 with size: 1.999512 MiB 00:05:49.978 element at address: 0x200000800000 with size: 1.996948 MiB 00:05:49.978 element at address: 0x20001bc00000 with size: 0.999878 MiB 00:05:49.978 element at address: 0x20001be00000 with size: 0.999878 MiB 00:05:49.978 element at address: 0x200034a00000 with size: 0.994446 MiB 00:05:49.978 element at address: 0x200009600000 with size: 0.959839 MiB 00:05:49.978 element at address: 0x200015e00000 with size: 0.954285 MiB 00:05:49.978 element at address: 0x20001c000000 with size: 0.936584 MiB 00:05:49.978 element at address: 0x200000200000 with size: 0.834839 MiB 00:05:49.978 element at address: 0x20001d800000 with size: 0.568237 MiB 00:05:49.978 element at address: 0x20000d800000 with size: 0.489075 MiB 00:05:49.978 element at address: 0x200003e00000 with size: 0.488831 MiB 00:05:49.978 element at address: 0x20001c200000 with size: 0.485657 MiB 00:05:49.978 element at address: 0x200007000000 with size: 0.480469 MiB 00:05:49.978 element at address: 0x20002ac00000 with size: 0.395752 MiB 00:05:49.978 element at address: 0x200003a00000 with size: 0.353027 MiB 00:05:49.978 list of standard malloc elements. size: 199.266052 MiB 00:05:49.978 element at address: 0x20000d9fff80 with size: 132.000122 MiB 00:05:49.978 element at address: 0x2000097fff80 with size: 64.000122 MiB 00:05:49.979 element at address: 0x20001bcfff80 with size: 1.000122 MiB 00:05:49.979 element at address: 0x20001befff80 with size: 1.000122 MiB 00:05:49.979 element at address: 0x20001c0fff80 with size: 1.000122 MiB 00:05:49.979 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:05:49.979 element at address: 0x20001c0eff00 with size: 0.062622 MiB 00:05:49.979 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:05:49.979 element at address: 0x20001c0efdc0 with size: 0.000305 MiB 00:05:49.979 element at address: 0x2000002d5b80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d5c40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d5d00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d5dc0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d5e80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d5f40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6000 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d60c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6180 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6240 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6300 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d63c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6480 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6540 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6600 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d66c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d68c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6980 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6a40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6b00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6bc0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6c80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6d40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6e00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6ec0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d6f80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7040 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7100 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d71c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7280 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7340 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7400 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d74c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7580 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7640 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7700 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d77c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7880 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7940 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7a00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7ac0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7b80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a5a600 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a5a800 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a5eac0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7ed80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7ee40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7ef00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7efc0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f080 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f140 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f200 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f2c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f380 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f440 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f500 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003a7f5c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003aff880 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003affa80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003affb40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d240 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d300 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d3c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d480 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d540 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d600 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d6c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d780 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d840 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d900 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7d9c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7da80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7db40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7dc00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7dcc0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7dd80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7de40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7df00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7dfc0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e080 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e140 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e200 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e2c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e380 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e440 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e500 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e5c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e680 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e740 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e800 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e8c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7e980 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7ea40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7eb00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7ebc0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7ec80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7ed40 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003e7ee00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200003eff0c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b000 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b0c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b180 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b240 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b300 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b3c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b480 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b540 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b600 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000707b6c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000070fb980 with size: 0.000183 MiB 00:05:49.979 element at address: 0x2000096fdd80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d340 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d400 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d4c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d580 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d640 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d700 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d7c0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d880 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87d940 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87da00 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d87dac0 with size: 0.000183 MiB 00:05:49.979 element at address: 0x20000d8fdd80 with size: 0.000183 MiB 00:05:49.979 element at address: 0x200015ef44c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001c0efc40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001c0efd00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001c2bc740 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891780 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891840 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891900 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8919c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891a80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891b40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891c00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891cc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891d80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891e40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891f00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d891fc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892080 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892140 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892200 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8922c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892380 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892440 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892500 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8925c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892680 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892740 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892800 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8928c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892980 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892a40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892b00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892bc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892c80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892d40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892e00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892ec0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d892f80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893040 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893100 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8931c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893280 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893340 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893400 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8934c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893580 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893640 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893700 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8937c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893880 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893940 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893a00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893ac0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893b80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893c40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893d00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893dc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893e80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d893f40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894000 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8940c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894180 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894240 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894300 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8943c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894480 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894540 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894600 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8946c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894780 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894840 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894900 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8949c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894a80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894b40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894c00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894cc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894d80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894e40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894f00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d894fc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d895080 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d895140 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d895200 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d8952c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d895380 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20001d895440 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac65500 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac655c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c1c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c3c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c480 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c540 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c600 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c6c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c780 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c840 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c900 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6c9c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6ca80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6cb40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6cc00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6ccc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6cd80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6ce40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6cf00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6cfc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d080 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d140 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d200 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d2c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d380 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d440 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d500 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d5c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d680 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d740 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d800 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d8c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6d980 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6da40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6db00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6dbc0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6dc80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6dd40 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6de00 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6dec0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6df80 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e040 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e100 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e1c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e280 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e340 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e400 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e4c0 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e580 with size: 0.000183 MiB 00:05:49.980 element at address: 0x20002ac6e640 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6e700 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6e7c0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6e880 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6e940 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6ea00 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6eac0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6eb80 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6ec40 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6ed00 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6edc0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6ee80 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6ef40 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f000 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f0c0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f180 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f240 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f300 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f3c0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f480 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f540 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f600 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f6c0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f780 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f840 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f900 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6f9c0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6fa80 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6fb40 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6fc00 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6fcc0 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6fd80 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6fe40 with size: 0.000183 MiB 00:05:49.981 element at address: 0x20002ac6ff00 with size: 0.000183 MiB 00:05:49.981 list of memzone associated elements. size: 646.796692 MiB 00:05:49.981 element at address: 0x20001d895500 with size: 211.416748 MiB 00:05:49.981 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:05:49.981 element at address: 0x20002ac6ffc0 with size: 157.562561 MiB 00:05:49.981 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:05:49.981 element at address: 0x200015ff4780 with size: 92.045044 MiB 00:05:49.981 associated memzone info: size: 92.044922 MiB name: MP_bdev_io_69700_0 00:05:49.981 element at address: 0x2000009ff380 with size: 48.003052 MiB 00:05:49.981 associated memzone info: size: 48.002930 MiB name: MP_evtpool_69700_0 00:05:49.981 element at address: 0x200003fff380 with size: 48.003052 MiB 00:05:49.981 associated memzone info: size: 48.002930 MiB name: MP_msgpool_69700_0 00:05:49.981 element at address: 0x2000071fdb80 with size: 36.008911 MiB 00:05:49.981 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_69700_0 00:05:49.981 element at address: 0x20001c3be940 with size: 20.255554 MiB 00:05:49.981 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:05:49.981 element at address: 0x200034bfeb40 with size: 18.005066 MiB 00:05:49.981 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:05:49.981 element at address: 0x2000005ffe00 with size: 2.000488 MiB 00:05:49.981 associated memzone info: size: 2.000366 MiB name: RG_MP_evtpool_69700 00:05:49.981 element at address: 0x200003bffe00 with size: 2.000488 MiB 00:05:49.981 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_69700 00:05:49.981 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:05:49.981 associated memzone info: size: 1.007996 MiB name: MP_evtpool_69700 00:05:49.981 element at address: 0x20000d8fde40 with size: 1.008118 MiB 00:05:49.981 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:05:49.981 element at address: 0x20001c2bc800 with size: 1.008118 MiB 00:05:49.981 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:05:49.981 element at address: 0x2000096fde40 with size: 1.008118 MiB 00:05:49.981 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:05:49.981 element at address: 0x2000070fba40 with size: 1.008118 MiB 00:05:49.981 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:05:49.981 element at address: 0x200003eff180 with size: 1.000488 MiB 00:05:49.981 associated memzone info: size: 1.000366 MiB name: RG_ring_0_69700 00:05:49.981 element at address: 0x200003affc00 with size: 1.000488 MiB 00:05:49.981 associated memzone info: size: 1.000366 MiB name: RG_ring_1_69700 00:05:49.981 element at address: 0x200015ef4580 with size: 1.000488 MiB 00:05:49.981 associated memzone info: size: 1.000366 MiB name: RG_ring_4_69700 00:05:49.981 element at address: 0x200034afe940 with size: 1.000488 MiB 00:05:49.981 associated memzone info: size: 1.000366 MiB name: RG_ring_5_69700 00:05:49.981 element at address: 0x200003a7f680 with size: 0.500488 MiB 00:05:49.981 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_69700 00:05:49.981 element at address: 0x200003e7eec0 with size: 0.500488 MiB 00:05:49.981 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_69700 00:05:49.981 element at address: 0x20000d87db80 with size: 0.500488 MiB 00:05:49.981 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:05:49.981 element at address: 0x20000707b780 with size: 0.500488 MiB 00:05:49.981 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:05:49.981 element at address: 0x20001c27c540 with size: 0.250488 MiB 00:05:49.981 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:05:49.981 element at address: 0x200003a5eb80 with size: 0.125488 MiB 00:05:49.981 associated memzone info: size: 0.125366 MiB name: RG_ring_2_69700 00:05:49.981 element at address: 0x2000096f5b80 with size: 0.031738 MiB 00:05:49.981 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:05:49.981 element at address: 0x20002ac65680 with size: 0.023743 MiB 00:05:49.981 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:05:49.981 element at address: 0x200003a5a8c0 with size: 0.016113 MiB 00:05:49.981 associated memzone info: size: 0.015991 MiB name: RG_ring_3_69700 00:05:49.981 element at address: 0x20002ac6b7c0 with size: 0.002441 MiB 00:05:49.981 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:05:49.981 element at address: 0x2000002d6780 with size: 0.000305 MiB 00:05:49.981 associated memzone info: size: 0.000183 MiB name: MP_msgpool_69700 00:05:49.981 element at address: 0x200003aff940 with size: 0.000305 MiB 00:05:49.981 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_69700 00:05:49.981 element at address: 0x200003a5a6c0 with size: 0.000305 MiB 00:05:49.981 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_69700 00:05:49.981 element at address: 0x20002ac6c280 with size: 0.000305 MiB 00:05:49.981 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:05:49.981 16:44:11 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:05:49.981 16:44:11 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 69700 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@950 -- # '[' -z 69700 ']' 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@954 -- # kill -0 69700 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@955 -- # uname 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69700 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69700' 00:05:49.981 killing process with pid 69700 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@969 -- # kill 69700 00:05:49.981 16:44:11 dpdk_mem_utility -- common/autotest_common.sh@974 -- # wait 69700 00:05:50.552 00:05:50.552 real 0m1.969s 00:05:50.552 user 0m1.729s 00:05:50.552 sys 0m0.675s 00:05:50.552 16:44:12 dpdk_mem_utility -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:50.552 16:44:12 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:05:50.552 ************************************ 00:05:50.552 END TEST dpdk_mem_utility 00:05:50.552 ************************************ 00:05:50.552 16:44:12 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:50.552 16:44:12 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:50.552 16:44:12 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:50.552 16:44:12 -- common/autotest_common.sh@10 -- # set +x 00:05:50.552 ************************************ 00:05:50.552 START TEST event 00:05:50.552 ************************************ 00:05:50.552 16:44:12 event -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:05:50.812 * Looking for test storage... 00:05:50.812 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:05:50.812 16:44:12 event -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:50.812 16:44:12 event -- common/autotest_common.sh@1681 -- # lcov --version 00:05:50.812 16:44:12 event -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:50.812 16:44:12 event -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:50.813 16:44:12 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:50.813 16:44:12 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:50.813 16:44:12 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:50.813 16:44:12 event -- scripts/common.sh@336 -- # IFS=.-: 00:05:50.813 16:44:12 event -- scripts/common.sh@336 -- # read -ra ver1 00:05:50.813 16:44:12 event -- scripts/common.sh@337 -- # IFS=.-: 00:05:50.813 16:44:12 event -- scripts/common.sh@337 -- # read -ra ver2 00:05:50.813 16:44:12 event -- scripts/common.sh@338 -- # local 'op=<' 00:05:50.813 16:44:12 event -- scripts/common.sh@340 -- # ver1_l=2 00:05:50.813 16:44:12 event -- scripts/common.sh@341 -- # ver2_l=1 00:05:50.813 16:44:12 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:50.813 16:44:12 event -- scripts/common.sh@344 -- # case "$op" in 00:05:50.813 16:44:12 event -- scripts/common.sh@345 -- # : 1 00:05:50.813 16:44:12 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:50.813 16:44:12 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:50.813 16:44:12 event -- scripts/common.sh@365 -- # decimal 1 00:05:50.813 16:44:12 event -- scripts/common.sh@353 -- # local d=1 00:05:50.813 16:44:12 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:50.813 16:44:12 event -- scripts/common.sh@355 -- # echo 1 00:05:50.813 16:44:12 event -- scripts/common.sh@365 -- # ver1[v]=1 00:05:50.813 16:44:12 event -- scripts/common.sh@366 -- # decimal 2 00:05:50.813 16:44:12 event -- scripts/common.sh@353 -- # local d=2 00:05:50.813 16:44:12 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:50.813 16:44:12 event -- scripts/common.sh@355 -- # echo 2 00:05:50.813 16:44:12 event -- scripts/common.sh@366 -- # ver2[v]=2 00:05:50.813 16:44:12 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:50.813 16:44:12 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:50.813 16:44:12 event -- scripts/common.sh@368 -- # return 0 00:05:50.813 16:44:12 event -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:50.813 16:44:12 event -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:50.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.813 --rc genhtml_branch_coverage=1 00:05:50.813 --rc genhtml_function_coverage=1 00:05:50.813 --rc genhtml_legend=1 00:05:50.813 --rc geninfo_all_blocks=1 00:05:50.813 --rc geninfo_unexecuted_blocks=1 00:05:50.813 00:05:50.813 ' 00:05:50.813 16:44:12 event -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:50.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.813 --rc genhtml_branch_coverage=1 00:05:50.813 --rc genhtml_function_coverage=1 00:05:50.813 --rc genhtml_legend=1 00:05:50.813 --rc geninfo_all_blocks=1 00:05:50.813 --rc geninfo_unexecuted_blocks=1 00:05:50.813 00:05:50.813 ' 00:05:50.813 16:44:12 event -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:50.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.813 --rc genhtml_branch_coverage=1 00:05:50.813 --rc genhtml_function_coverage=1 00:05:50.813 --rc genhtml_legend=1 00:05:50.813 --rc geninfo_all_blocks=1 00:05:50.813 --rc geninfo_unexecuted_blocks=1 00:05:50.813 00:05:50.813 ' 00:05:50.813 16:44:12 event -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:50.813 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:50.813 --rc genhtml_branch_coverage=1 00:05:50.813 --rc genhtml_function_coverage=1 00:05:50.813 --rc genhtml_legend=1 00:05:50.813 --rc geninfo_all_blocks=1 00:05:50.813 --rc geninfo_unexecuted_blocks=1 00:05:50.813 00:05:50.813 ' 00:05:50.813 16:44:12 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:05:50.813 16:44:12 event -- bdev/nbd_common.sh@6 -- # set -e 00:05:50.813 16:44:12 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:50.813 16:44:12 event -- common/autotest_common.sh@1101 -- # '[' 6 -le 1 ']' 00:05:50.813 16:44:12 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:50.813 16:44:12 event -- common/autotest_common.sh@10 -- # set +x 00:05:50.813 ************************************ 00:05:50.813 START TEST event_perf 00:05:50.813 ************************************ 00:05:50.813 16:44:12 event.event_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:05:51.073 Running I/O for 1 seconds...[2024-09-29 16:44:12.501030] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:51.073 [2024-09-29 16:44:12.501183] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69786 ] 00:05:51.073 [2024-09-29 16:44:12.639392] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:51.073 [2024-09-29 16:44:12.718418] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:51.073 [2024-09-29 16:44:12.718598] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:05:51.073 Running I/O for 1 seconds...[2024-09-29 16:44:12.718827] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:05:51.073 [2024-09-29 16:44:12.718681] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:52.455 00:05:52.455 lcore 0: 98887 00:05:52.455 lcore 1: 98890 00:05:52.455 lcore 2: 98893 00:05:52.455 lcore 3: 98891 00:05:52.455 done. 00:05:52.455 00:05:52.455 real 0m1.401s 00:05:52.455 user 0m4.154s 00:05:52.455 sys 0m0.122s 00:05:52.455 16:44:13 event.event_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:52.455 16:44:13 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:05:52.455 ************************************ 00:05:52.455 END TEST event_perf 00:05:52.455 ************************************ 00:05:52.455 16:44:13 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:52.455 16:44:13 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:05:52.455 16:44:13 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:52.455 16:44:13 event -- common/autotest_common.sh@10 -- # set +x 00:05:52.455 ************************************ 00:05:52.455 START TEST event_reactor 00:05:52.455 ************************************ 00:05:52.455 16:44:13 event.event_reactor -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:05:52.455 [2024-09-29 16:44:13.978350] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:52.455 [2024-09-29 16:44:13.978540] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69820 ] 00:05:52.455 [2024-09-29 16:44:14.125548] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:52.715 [2024-09-29 16:44:14.219543] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:54.097 test_start 00:05:54.097 oneshot 00:05:54.097 tick 100 00:05:54.097 tick 100 00:05:54.097 tick 250 00:05:54.097 tick 100 00:05:54.097 tick 100 00:05:54.097 tick 100 00:05:54.097 tick 250 00:05:54.097 tick 500 00:05:54.097 tick 100 00:05:54.097 tick 100 00:05:54.097 tick 250 00:05:54.097 tick 100 00:05:54.097 tick 100 00:05:54.097 test_end 00:05:54.097 00:05:54.097 real 0m1.421s 00:05:54.097 user 0m1.193s 00:05:54.097 sys 0m0.116s 00:05:54.097 16:44:15 event.event_reactor -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:54.097 16:44:15 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:05:54.097 ************************************ 00:05:54.097 END TEST event_reactor 00:05:54.097 ************************************ 00:05:54.097 16:44:15 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:54.097 16:44:15 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:05:54.097 16:44:15 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:54.097 16:44:15 event -- common/autotest_common.sh@10 -- # set +x 00:05:54.097 ************************************ 00:05:54.097 START TEST event_reactor_perf 00:05:54.097 ************************************ 00:05:54.097 16:44:15 event.event_reactor_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:05:54.097 [2024-09-29 16:44:15.471026] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:54.097 [2024-09-29 16:44:15.471133] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69862 ] 00:05:54.097 [2024-09-29 16:44:15.615936] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:54.097 [2024-09-29 16:44:15.699317] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:55.478 test_start 00:05:55.478 test_end 00:05:55.478 Performance: 376485 events per second 00:05:55.478 00:05:55.478 real 0m1.406s 00:05:55.478 user 0m1.177s 00:05:55.478 sys 0m0.118s 00:05:55.478 ************************************ 00:05:55.478 END TEST event_reactor_perf 00:05:55.478 ************************************ 00:05:55.478 16:44:16 event.event_reactor_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:55.478 16:44:16 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:05:55.478 16:44:16 event -- event/event.sh@49 -- # uname -s 00:05:55.478 16:44:16 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:05:55.478 16:44:16 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:55.478 16:44:16 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:55.478 16:44:16 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:55.478 16:44:16 event -- common/autotest_common.sh@10 -- # set +x 00:05:55.478 ************************************ 00:05:55.478 START TEST event_scheduler 00:05:55.478 ************************************ 00:05:55.478 16:44:16 event.event_scheduler -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:05:55.478 * Looking for test storage... 00:05:55.478 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1681 -- # lcov --version 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:55.478 16:44:17 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:55.478 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.478 --rc genhtml_branch_coverage=1 00:05:55.478 --rc genhtml_function_coverage=1 00:05:55.478 --rc genhtml_legend=1 00:05:55.478 --rc geninfo_all_blocks=1 00:05:55.478 --rc geninfo_unexecuted_blocks=1 00:05:55.478 00:05:55.478 ' 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:55.478 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.478 --rc genhtml_branch_coverage=1 00:05:55.478 --rc genhtml_function_coverage=1 00:05:55.478 --rc genhtml_legend=1 00:05:55.478 --rc geninfo_all_blocks=1 00:05:55.478 --rc geninfo_unexecuted_blocks=1 00:05:55.478 00:05:55.478 ' 00:05:55.478 16:44:17 event.event_scheduler -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:55.478 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.478 --rc genhtml_branch_coverage=1 00:05:55.478 --rc genhtml_function_coverage=1 00:05:55.478 --rc genhtml_legend=1 00:05:55.478 --rc geninfo_all_blocks=1 00:05:55.478 --rc geninfo_unexecuted_blocks=1 00:05:55.478 00:05:55.479 ' 00:05:55.479 16:44:17 event.event_scheduler -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:55.479 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:55.479 --rc genhtml_branch_coverage=1 00:05:55.479 --rc genhtml_function_coverage=1 00:05:55.479 --rc genhtml_legend=1 00:05:55.479 --rc geninfo_all_blocks=1 00:05:55.479 --rc geninfo_unexecuted_blocks=1 00:05:55.479 00:05:55.479 ' 00:05:55.479 16:44:17 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:05:55.479 16:44:17 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=69928 00:05:55.479 16:44:17 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:05:55.479 16:44:17 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:05:55.479 16:44:17 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 69928 00:05:55.479 16:44:17 event.event_scheduler -- common/autotest_common.sh@831 -- # '[' -z 69928 ']' 00:05:55.479 16:44:17 event.event_scheduler -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:55.479 16:44:17 event.event_scheduler -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:55.738 16:44:17 event.event_scheduler -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:55.738 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:55.738 16:44:17 event.event_scheduler -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:55.738 16:44:17 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:55.738 [2024-09-29 16:44:17.222188] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:05:55.738 [2024-09-29 16:44:17.222321] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69928 ] 00:05:55.738 [2024-09-29 16:44:17.367296] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:05:55.996 [2024-09-29 16:44:17.416200] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:55.996 [2024-09-29 16:44:17.416345] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:05:55.996 [2024-09-29 16:44:17.416483] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:05:55.996 [2024-09-29 16:44:17.416555] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@864 -- # return 0 00:05:56.563 16:44:18 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:56.563 POWER: Cannot set governor of lcore 0 to userspace 00:05:56.563 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:56.563 POWER: Cannot set governor of lcore 0 to performance 00:05:56.563 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:05:56.563 POWER: Cannot set governor of lcore 0 to userspace 00:05:56.563 GUEST_CHANNEL: Unable to to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:05:56.563 POWER: Unable to set Power Management Environment for lcore 0 00:05:56.563 [2024-09-29 16:44:18.057446] dpdk_governor.c: 130:_init_core: *ERROR*: Failed to initialize on core0 00:05:56.563 [2024-09-29 16:44:18.057488] dpdk_governor.c: 191:_init: *ERROR*: Failed to initialize on core0 00:05:56.563 [2024-09-29 16:44:18.057510] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:05:56.563 [2024-09-29 16:44:18.057552] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:05:56.563 [2024-09-29 16:44:18.057561] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:05:56.563 [2024-09-29 16:44:18.057571] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 [2024-09-29 16:44:18.127630] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 ************************************ 00:05:56.563 START TEST scheduler_create_thread 00:05:56.563 ************************************ 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1125 -- # scheduler_create_thread 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 2 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 3 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 4 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 5 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 6 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 7 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:56.563 8 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.563 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:57.132 9 00:05:57.132 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:57.132 16:44:18 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:05:57.132 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:57.132 16:44:18 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:58.510 10 00:05:58.510 16:44:19 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:58.510 16:44:19 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:05:58.510 16:44:19 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:58.510 16:44:19 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:59.078 16:44:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:59.078 16:44:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:05:59.078 16:44:20 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:05:59.078 16:44:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:59.078 16:44:20 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:00.016 16:44:21 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:00.016 16:44:21 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:06:00.016 16:44:21 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:00.016 16:44:21 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:00.584 16:44:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:00.584 16:44:22 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:06:00.584 16:44:22 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:06:00.584 16:44:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:00.584 16:44:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:01.151 16:44:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:01.151 00:06:01.151 real 0m4.598s 00:06:01.151 user 0m0.029s 00:06:01.151 sys 0m0.007s 00:06:01.151 ************************************ 00:06:01.151 END TEST scheduler_create_thread 00:06:01.151 ************************************ 00:06:01.151 16:44:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:01.151 16:44:22 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:01.151 16:44:22 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:06:01.151 16:44:22 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 69928 00:06:01.151 16:44:22 event.event_scheduler -- common/autotest_common.sh@950 -- # '[' -z 69928 ']' 00:06:01.151 16:44:22 event.event_scheduler -- common/autotest_common.sh@954 -- # kill -0 69928 00:06:01.151 16:44:22 event.event_scheduler -- common/autotest_common.sh@955 -- # uname 00:06:01.151 16:44:22 event.event_scheduler -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:01.151 16:44:22 event.event_scheduler -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69928 00:06:01.409 killing process with pid 69928 00:06:01.409 16:44:22 event.event_scheduler -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:06:01.409 16:44:22 event.event_scheduler -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:06:01.409 16:44:22 event.event_scheduler -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69928' 00:06:01.409 16:44:22 event.event_scheduler -- common/autotest_common.sh@969 -- # kill 69928 00:06:01.409 16:44:22 event.event_scheduler -- common/autotest_common.sh@974 -- # wait 69928 00:06:01.409 [2024-09-29 16:44:23.017579] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:06:01.700 ************************************ 00:06:01.700 END TEST event_scheduler 00:06:01.700 ************************************ 00:06:01.700 00:06:01.700 real 0m6.410s 00:06:01.700 user 0m13.821s 00:06:01.700 sys 0m0.500s 00:06:01.700 16:44:23 event.event_scheduler -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:01.700 16:44:23 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:01.960 16:44:23 event -- event/event.sh@51 -- # modprobe -n nbd 00:06:01.960 16:44:23 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:06:01.960 16:44:23 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:01.960 16:44:23 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:01.960 16:44:23 event -- common/autotest_common.sh@10 -- # set +x 00:06:01.960 ************************************ 00:06:01.960 START TEST app_repeat 00:06:01.960 ************************************ 00:06:01.960 16:44:23 event.app_repeat -- common/autotest_common.sh@1125 -- # app_repeat_test 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70050 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70050' 00:06:01.960 Process app_repeat pid: 70050 00:06:01.960 spdk_app_start Round 0 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:06:01.960 16:44:23 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70050 /var/tmp/spdk-nbd.sock 00:06:01.960 16:44:23 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70050 ']' 00:06:01.960 16:44:23 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:01.960 16:44:23 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:01.960 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:01.960 16:44:23 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:01.960 16:44:23 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:01.960 16:44:23 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:01.960 [2024-09-29 16:44:23.460688] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:01.960 [2024-09-29 16:44:23.460812] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70050 ] 00:06:01.960 [2024-09-29 16:44:23.606658] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:02.219 [2024-09-29 16:44:23.682829] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:02.219 [2024-09-29 16:44:23.682932] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:02.784 16:44:24 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:02.784 16:44:24 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:02.784 16:44:24 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:03.043 Malloc0 00:06:03.043 16:44:24 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:03.301 Malloc1 00:06:03.301 16:44:24 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:03.301 /dev/nbd0 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:03.301 16:44:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:03.301 16:44:24 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:03.301 16:44:24 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:03.301 16:44:24 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:03.302 16:44:24 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:03.302 16:44:24 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:03.302 16:44:24 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:03.302 16:44:24 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:03.302 16:44:24 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:03.302 16:44:24 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:03.302 1+0 records in 00:06:03.302 1+0 records out 00:06:03.302 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000502396 s, 8.2 MB/s 00:06:03.302 16:44:24 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:03.559 16:44:24 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:03.559 16:44:24 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:03.559 16:44:24 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:03.559 16:44:24 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:03.559 16:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:03.559 16:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:03.559 16:44:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:03.559 /dev/nbd1 00:06:03.559 16:44:25 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:03.559 16:44:25 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:03.559 1+0 records in 00:06:03.559 1+0 records out 00:06:03.559 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000312361 s, 13.1 MB/s 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:03.559 16:44:25 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:03.559 16:44:25 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:03.559 16:44:25 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:03.559 16:44:25 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:03.559 16:44:25 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:03.559 16:44:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:03.817 { 00:06:03.817 "nbd_device": "/dev/nbd0", 00:06:03.817 "bdev_name": "Malloc0" 00:06:03.817 }, 00:06:03.817 { 00:06:03.817 "nbd_device": "/dev/nbd1", 00:06:03.817 "bdev_name": "Malloc1" 00:06:03.817 } 00:06:03.817 ]' 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:03.817 { 00:06:03.817 "nbd_device": "/dev/nbd0", 00:06:03.817 "bdev_name": "Malloc0" 00:06:03.817 }, 00:06:03.817 { 00:06:03.817 "nbd_device": "/dev/nbd1", 00:06:03.817 "bdev_name": "Malloc1" 00:06:03.817 } 00:06:03.817 ]' 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:03.817 /dev/nbd1' 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:03.817 /dev/nbd1' 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:03.817 256+0 records in 00:06:03.817 256+0 records out 00:06:03.817 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0124548 s, 84.2 MB/s 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:03.817 16:44:25 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:04.100 256+0 records in 00:06:04.100 256+0 records out 00:06:04.100 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0231691 s, 45.3 MB/s 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:04.100 256+0 records in 00:06:04.100 256+0 records out 00:06:04.100 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0226692 s, 46.3 MB/s 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:04.100 16:44:25 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:04.360 16:44:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:04.619 16:44:26 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:04.619 16:44:26 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:04.878 16:44:26 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:05.136 [2024-09-29 16:44:26.601607] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:05.137 [2024-09-29 16:44:26.643859] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:05.137 [2024-09-29 16:44:26.643865] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:05.137 [2024-09-29 16:44:26.685787] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:05.137 [2024-09-29 16:44:26.685862] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:08.420 spdk_app_start Round 1 00:06:08.420 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:08.420 16:44:29 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:08.420 16:44:29 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:06:08.420 16:44:29 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70050 /var/tmp/spdk-nbd.sock 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70050 ']' 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:08.420 16:44:29 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:08.420 16:44:29 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:08.420 Malloc0 00:06:08.420 16:44:29 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:08.420 Malloc1 00:06:08.420 16:44:30 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:08.420 16:44:30 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:08.679 /dev/nbd0 00:06:08.679 16:44:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:08.679 16:44:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:08.679 1+0 records in 00:06:08.679 1+0 records out 00:06:08.679 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000430804 s, 9.5 MB/s 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:08.679 16:44:30 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:08.679 16:44:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:08.679 16:44:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:08.679 16:44:30 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:08.938 /dev/nbd1 00:06:08.938 16:44:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:08.938 16:44:30 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:08.938 16:44:30 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:08.939 16:44:30 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:08.939 1+0 records in 00:06:08.939 1+0 records out 00:06:08.939 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000322244 s, 12.7 MB/s 00:06:08.939 16:44:30 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:08.939 16:44:30 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:08.939 16:44:30 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:08.939 16:44:30 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:08.939 16:44:30 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:08.939 16:44:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:08.939 16:44:30 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:08.939 16:44:30 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:08.939 16:44:30 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:08.939 16:44:30 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:09.198 16:44:30 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:09.198 { 00:06:09.198 "nbd_device": "/dev/nbd0", 00:06:09.199 "bdev_name": "Malloc0" 00:06:09.199 }, 00:06:09.199 { 00:06:09.199 "nbd_device": "/dev/nbd1", 00:06:09.199 "bdev_name": "Malloc1" 00:06:09.199 } 00:06:09.199 ]' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:09.199 { 00:06:09.199 "nbd_device": "/dev/nbd0", 00:06:09.199 "bdev_name": "Malloc0" 00:06:09.199 }, 00:06:09.199 { 00:06:09.199 "nbd_device": "/dev/nbd1", 00:06:09.199 "bdev_name": "Malloc1" 00:06:09.199 } 00:06:09.199 ]' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:09.199 /dev/nbd1' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:09.199 /dev/nbd1' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:09.199 256+0 records in 00:06:09.199 256+0 records out 00:06:09.199 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00362464 s, 289 MB/s 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:09.199 256+0 records in 00:06:09.199 256+0 records out 00:06:09.199 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0235683 s, 44.5 MB/s 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:09.199 256+0 records in 00:06:09.199 256+0 records out 00:06:09.199 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0237852 s, 44.1 MB/s 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:09.199 16:44:30 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:09.458 16:44:30 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:09.458 16:44:30 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:09.458 16:44:30 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:09.458 16:44:30 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:09.458 16:44:30 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:09.458 16:44:30 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:09.458 16:44:30 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:09.458 16:44:31 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:09.717 16:44:31 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:09.975 16:44:31 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:09.975 16:44:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:09.975 16:44:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:09.975 16:44:31 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:09.975 16:44:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:09.976 16:44:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:09.976 16:44:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:09.976 16:44:31 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:09.976 16:44:31 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:09.976 16:44:31 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:09.976 16:44:31 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:09.976 16:44:31 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:09.976 16:44:31 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:10.235 16:44:31 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:10.493 [2024-09-29 16:44:31.913815] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:10.493 [2024-09-29 16:44:31.956154] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:10.493 [2024-09-29 16:44:31.956186] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:10.493 [2024-09-29 16:44:31.998071] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:10.493 [2024-09-29 16:44:31.998131] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:13.777 spdk_app_start Round 2 00:06:13.777 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:13.777 16:44:34 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:13.777 16:44:34 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:06:13.777 16:44:34 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70050 /var/tmp/spdk-nbd.sock 00:06:13.777 16:44:34 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70050 ']' 00:06:13.777 16:44:34 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:13.777 16:44:34 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:13.777 16:44:34 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:13.777 16:44:34 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:13.777 16:44:34 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:13.777 16:44:34 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:13.778 16:44:34 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:13.778 16:44:34 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:13.778 Malloc0 00:06:13.778 16:44:35 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:13.778 Malloc1 00:06:13.778 16:44:35 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:13.778 16:44:35 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:14.037 /dev/nbd0 00:06:14.037 16:44:35 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:14.037 16:44:35 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:14.037 1+0 records in 00:06:14.037 1+0 records out 00:06:14.037 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000455577 s, 9.0 MB/s 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:14.037 16:44:35 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:14.037 16:44:35 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:14.037 16:44:35 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:14.037 16:44:35 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:14.296 /dev/nbd1 00:06:14.296 16:44:35 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:14.296 16:44:35 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:14.296 1+0 records in 00:06:14.296 1+0 records out 00:06:14.296 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000256476 s, 16.0 MB/s 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:14.296 16:44:35 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:14.296 16:44:35 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:14.296 16:44:35 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:14.296 16:44:35 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:14.296 16:44:35 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:14.296 16:44:35 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:14.555 { 00:06:14.555 "nbd_device": "/dev/nbd0", 00:06:14.555 "bdev_name": "Malloc0" 00:06:14.555 }, 00:06:14.555 { 00:06:14.555 "nbd_device": "/dev/nbd1", 00:06:14.555 "bdev_name": "Malloc1" 00:06:14.555 } 00:06:14.555 ]' 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:14.555 { 00:06:14.555 "nbd_device": "/dev/nbd0", 00:06:14.555 "bdev_name": "Malloc0" 00:06:14.555 }, 00:06:14.555 { 00:06:14.555 "nbd_device": "/dev/nbd1", 00:06:14.555 "bdev_name": "Malloc1" 00:06:14.555 } 00:06:14.555 ]' 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:14.555 /dev/nbd1' 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:14.555 /dev/nbd1' 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:14.555 256+0 records in 00:06:14.555 256+0 records out 00:06:14.555 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0126568 s, 82.8 MB/s 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:14.555 256+0 records in 00:06:14.555 256+0 records out 00:06:14.555 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0162801 s, 64.4 MB/s 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:14.555 16:44:36 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:14.555 256+0 records in 00:06:14.555 256+0 records out 00:06:14.556 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0215915 s, 48.6 MB/s 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:14.556 16:44:36 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:14.815 16:44:36 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:15.074 16:44:36 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:15.333 16:44:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:15.333 16:44:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:15.333 16:44:36 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:15.333 16:44:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:15.333 16:44:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:15.334 16:44:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:15.334 16:44:36 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:15.334 16:44:36 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:15.334 16:44:36 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:15.334 16:44:36 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:15.334 16:44:36 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:15.334 16:44:36 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:15.593 16:44:37 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:15.593 [2024-09-29 16:44:37.174446] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:15.593 [2024-09-29 16:44:37.217018] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:15.593 [2024-09-29 16:44:37.217025] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:15.593 [2024-09-29 16:44:37.260092] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:15.593 [2024-09-29 16:44:37.260145] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:18.877 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:18.877 16:44:40 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70050 /var/tmp/spdk-nbd.sock 00:06:18.877 16:44:40 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70050 ']' 00:06:18.877 16:44:40 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:18.877 16:44:40 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:18.877 16:44:40 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:18.877 16:44:40 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:18.877 16:44:40 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:18.877 16:44:40 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:18.878 16:44:40 event.app_repeat -- event/event.sh@39 -- # killprocess 70050 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@950 -- # '[' -z 70050 ']' 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@954 -- # kill -0 70050 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@955 -- # uname 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70050 00:06:18.878 killing process with pid 70050 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70050' 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@969 -- # kill 70050 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@974 -- # wait 70050 00:06:18.878 spdk_app_start is called in Round 0. 00:06:18.878 Shutdown signal received, stop current app iteration 00:06:18.878 Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 reinitialization... 00:06:18.878 spdk_app_start is called in Round 1. 00:06:18.878 Shutdown signal received, stop current app iteration 00:06:18.878 Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 reinitialization... 00:06:18.878 spdk_app_start is called in Round 2. 00:06:18.878 Shutdown signal received, stop current app iteration 00:06:18.878 Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 reinitialization... 00:06:18.878 spdk_app_start is called in Round 3. 00:06:18.878 Shutdown signal received, stop current app iteration 00:06:18.878 ************************************ 00:06:18.878 END TEST app_repeat 00:06:18.878 ************************************ 00:06:18.878 16:44:40 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:06:18.878 16:44:40 event.app_repeat -- event/event.sh@42 -- # return 0 00:06:18.878 00:06:18.878 real 0m17.047s 00:06:18.878 user 0m37.217s 00:06:18.878 sys 0m2.614s 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:18.878 16:44:40 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:18.878 16:44:40 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:06:18.878 16:44:40 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:18.878 16:44:40 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:18.878 16:44:40 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:18.878 16:44:40 event -- common/autotest_common.sh@10 -- # set +x 00:06:18.878 ************************************ 00:06:18.878 START TEST cpu_locks 00:06:18.878 ************************************ 00:06:18.878 16:44:40 event.cpu_locks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:19.138 * Looking for test storage... 00:06:19.138 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1681 -- # lcov --version 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:19.138 16:44:40 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:19.138 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.138 --rc genhtml_branch_coverage=1 00:06:19.138 --rc genhtml_function_coverage=1 00:06:19.138 --rc genhtml_legend=1 00:06:19.138 --rc geninfo_all_blocks=1 00:06:19.138 --rc geninfo_unexecuted_blocks=1 00:06:19.138 00:06:19.138 ' 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:19.138 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.138 --rc genhtml_branch_coverage=1 00:06:19.138 --rc genhtml_function_coverage=1 00:06:19.138 --rc genhtml_legend=1 00:06:19.138 --rc geninfo_all_blocks=1 00:06:19.138 --rc geninfo_unexecuted_blocks=1 00:06:19.138 00:06:19.138 ' 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:19.138 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.138 --rc genhtml_branch_coverage=1 00:06:19.138 --rc genhtml_function_coverage=1 00:06:19.138 --rc genhtml_legend=1 00:06:19.138 --rc geninfo_all_blocks=1 00:06:19.138 --rc geninfo_unexecuted_blocks=1 00:06:19.138 00:06:19.138 ' 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:19.138 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.138 --rc genhtml_branch_coverage=1 00:06:19.138 --rc genhtml_function_coverage=1 00:06:19.138 --rc genhtml_legend=1 00:06:19.138 --rc geninfo_all_blocks=1 00:06:19.138 --rc geninfo_unexecuted_blocks=1 00:06:19.138 00:06:19.138 ' 00:06:19.138 16:44:40 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:06:19.138 16:44:40 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:06:19.138 16:44:40 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:06:19.138 16:44:40 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:19.138 16:44:40 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:19.138 ************************************ 00:06:19.138 START TEST default_locks 00:06:19.138 ************************************ 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- common/autotest_common.sh@1125 -- # default_locks 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70469 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70469 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 70469 ']' 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:19.138 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:19.138 16:44:40 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:19.398 [2024-09-29 16:44:40.833895] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:19.398 [2024-09-29 16:44:40.834102] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70469 ] 00:06:19.398 [2024-09-29 16:44:40.977942] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:19.398 [2024-09-29 16:44:41.023348] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:20.041 16:44:41 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:20.041 16:44:41 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 0 00:06:20.041 16:44:41 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70469 00:06:20.041 16:44:41 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70469 00:06:20.041 16:44:41 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70469 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@950 -- # '[' -z 70469 ']' 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # kill -0 70469 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # uname 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70469 00:06:20.611 killing process with pid 70469 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70469' 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@969 -- # kill 70469 00:06:20.611 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@974 -- # wait 70469 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70469 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # local es=0 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 70469 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # waitforlisten 70469 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 70469 ']' 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:20.870 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:20.870 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (70469) - No such process 00:06:20.870 ERROR: process (pid: 70469) is no longer running 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 1 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # es=1 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:20.870 00:06:20.870 real 0m1.721s 00:06:20.870 user 0m1.708s 00:06:20.870 sys 0m0.582s 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:20.870 ************************************ 00:06:20.870 END TEST default_locks 00:06:20.870 ************************************ 00:06:20.870 16:44:42 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:20.870 16:44:42 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:06:20.870 16:44:42 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:20.870 16:44:42 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:20.870 16:44:42 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:20.870 ************************************ 00:06:20.870 START TEST default_locks_via_rpc 00:06:20.870 ************************************ 00:06:20.870 16:44:42 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1125 -- # default_locks_via_rpc 00:06:20.870 16:44:42 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70521 00:06:20.870 16:44:42 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:20.870 16:44:42 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70521 00:06:20.870 16:44:42 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70521 ']' 00:06:20.870 16:44:42 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:20.871 16:44:42 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:20.871 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:20.871 16:44:42 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:20.871 16:44:42 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:20.871 16:44:42 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:21.129 [2024-09-29 16:44:42.623519] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:21.129 [2024-09-29 16:44:42.623670] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70521 ] 00:06:21.129 [2024-09-29 16:44:42.768270] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:21.395 [2024-09-29 16:44:42.813973] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70521 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70521 00:06:21.964 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70521 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@950 -- # '[' -z 70521 ']' 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # kill -0 70521 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # uname 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70521 00:06:22.224 killing process with pid 70521 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70521' 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@969 -- # kill 70521 00:06:22.224 16:44:43 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@974 -- # wait 70521 00:06:22.792 00:06:22.792 real 0m1.670s 00:06:22.792 user 0m1.637s 00:06:22.792 sys 0m0.558s 00:06:22.792 ************************************ 00:06:22.792 END TEST default_locks_via_rpc 00:06:22.792 ************************************ 00:06:22.792 16:44:44 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:22.792 16:44:44 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:22.792 16:44:44 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:06:22.792 16:44:44 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:22.792 16:44:44 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:22.792 16:44:44 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:22.792 ************************************ 00:06:22.792 START TEST non_locking_app_on_locked_coremask 00:06:22.792 ************************************ 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # non_locking_app_on_locked_coremask 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=70569 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 70569 /var/tmp/spdk.sock 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70569 ']' 00:06:22.792 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:22.792 16:44:44 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:22.792 [2024-09-29 16:44:44.359045] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:22.792 [2024-09-29 16:44:44.359293] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70569 ] 00:06:23.050 [2024-09-29 16:44:44.489150] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:23.050 [2024-09-29 16:44:44.533049] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=70585 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 70585 /var/tmp/spdk2.sock 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70585 ']' 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:23.617 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:23.617 16:44:45 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:23.617 [2024-09-29 16:44:45.248571] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:23.617 [2024-09-29 16:44:45.248798] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70585 ] 00:06:23.876 [2024-09-29 16:44:45.380708] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:23.876 [2024-09-29 16:44:45.380778] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:23.876 [2024-09-29 16:44:45.470475] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:24.443 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:24.443 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:24.443 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 70569 00:06:24.443 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70569 00:06:24.443 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:25.379 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 70569 00:06:25.379 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70569 ']' 00:06:25.379 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 70569 00:06:25.379 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:25.379 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:25.379 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70569 00:06:25.379 killing process with pid 70569 00:06:25.379 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:25.380 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:25.380 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70569' 00:06:25.380 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 70569 00:06:25.380 16:44:46 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 70569 00:06:25.947 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 70585 00:06:25.947 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70585 ']' 00:06:25.947 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 70585 00:06:25.947 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:25.947 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:25.947 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70585 00:06:26.206 killing process with pid 70585 00:06:26.206 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:26.206 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:26.206 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70585' 00:06:26.206 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 70585 00:06:26.206 16:44:47 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 70585 00:06:26.465 ************************************ 00:06:26.465 END TEST non_locking_app_on_locked_coremask 00:06:26.465 ************************************ 00:06:26.465 00:06:26.465 real 0m3.743s 00:06:26.465 user 0m3.880s 00:06:26.465 sys 0m1.140s 00:06:26.465 16:44:48 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:26.465 16:44:48 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:26.465 16:44:48 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:06:26.465 16:44:48 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:26.465 16:44:48 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:26.465 16:44:48 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:26.465 ************************************ 00:06:26.465 START TEST locking_app_on_unlocked_coremask 00:06:26.465 ************************************ 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_unlocked_coremask 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=70654 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 70654 /var/tmp/spdk.sock 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70654 ']' 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:26.465 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:26.465 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:26.724 [2024-09-29 16:44:48.171053] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:26.724 [2024-09-29 16:44:48.171305] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70654 ] 00:06:26.724 [2024-09-29 16:44:48.315728] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:26.724 [2024-09-29 16:44:48.315826] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:26.724 [2024-09-29 16:44:48.361013] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=70664 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 70664 /var/tmp/spdk2.sock 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70664 ']' 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:27.662 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:27.662 16:44:48 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:27.662 [2024-09-29 16:44:49.042168] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:27.662 [2024-09-29 16:44:49.042355] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70664 ] 00:06:27.662 [2024-09-29 16:44:49.176917] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:27.662 [2024-09-29 16:44:49.270184] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:28.232 16:44:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:28.232 16:44:49 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:28.232 16:44:49 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 70664 00:06:28.232 16:44:49 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70664 00:06:28.232 16:44:49 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 70654 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70654 ']' 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 70654 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70654 00:06:29.170 killing process with pid 70654 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70654' 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 70654 00:06:29.170 16:44:50 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 70654 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 70664 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70664 ']' 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 70664 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70664 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:30.111 killing process with pid 70664 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70664' 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 70664 00:06:30.111 16:44:51 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 70664 00:06:30.370 00:06:30.371 real 0m3.930s 00:06:30.371 user 0m4.081s 00:06:30.371 sys 0m1.250s 00:06:30.371 16:44:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:30.371 16:44:52 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:30.371 ************************************ 00:06:30.371 END TEST locking_app_on_unlocked_coremask 00:06:30.371 ************************************ 00:06:30.631 16:44:52 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:06:30.631 16:44:52 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:30.631 16:44:52 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:30.631 16:44:52 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:30.631 ************************************ 00:06:30.631 START TEST locking_app_on_locked_coremask 00:06:30.631 ************************************ 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_locked_coremask 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=70735 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 70735 /var/tmp/spdk.sock 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70735 ']' 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:30.631 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:30.631 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:30.631 [2024-09-29 16:44:52.159047] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:30.631 [2024-09-29 16:44:52.159268] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70735 ] 00:06:30.631 [2024-09-29 16:44:52.301899] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:30.891 [2024-09-29 16:44:52.347113] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=70746 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 70746 /var/tmp/spdk2.sock 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # local es=0 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 70746 /var/tmp/spdk2.sock 00:06:31.459 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # waitforlisten 70746 /var/tmp/spdk2.sock 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70746 ']' 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:31.460 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:31.460 16:44:52 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:31.460 [2024-09-29 16:44:53.023309] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:31.460 [2024-09-29 16:44:53.023867] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70746 ] 00:06:31.720 [2024-09-29 16:44:53.159902] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 70735 has claimed it. 00:06:31.720 [2024-09-29 16:44:53.160004] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:32.289 ERROR: process (pid: 70746) is no longer running 00:06:32.289 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (70746) - No such process 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 1 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # es=1 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 70735 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70735 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 70735 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70735 ']' 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 70735 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:32.289 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70735 00:06:32.549 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:32.549 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:32.549 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70735' 00:06:32.549 killing process with pid 70735 00:06:32.549 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 70735 00:06:32.549 16:44:53 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 70735 00:06:32.809 00:06:32.809 real 0m2.284s 00:06:32.809 user 0m2.437s 00:06:32.809 sys 0m0.634s 00:06:32.809 16:44:54 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:32.809 16:44:54 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:32.809 ************************************ 00:06:32.809 END TEST locking_app_on_locked_coremask 00:06:32.809 ************************************ 00:06:32.809 16:44:54 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:06:32.809 16:44:54 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:32.809 16:44:54 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:32.809 16:44:54 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:32.809 ************************************ 00:06:32.809 START TEST locking_overlapped_coremask 00:06:32.809 ************************************ 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=70799 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 70799 /var/tmp/spdk.sock 00:06:32.809 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 70799 ']' 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:32.809 16:44:54 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:33.069 [2024-09-29 16:44:54.516661] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:33.069 [2024-09-29 16:44:54.516836] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70799 ] 00:06:33.069 [2024-09-29 16:44:54.662855] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:33.069 [2024-09-29 16:44:54.710319] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:33.069 [2024-09-29 16:44:54.710412] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:33.069 [2024-09-29 16:44:54.710523] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:34.028 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:34.028 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:34.028 16:44:55 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:06:34.028 16:44:55 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=70817 00:06:34.028 16:44:55 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 70817 /var/tmp/spdk2.sock 00:06:34.028 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # local es=0 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 70817 /var/tmp/spdk2.sock 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # waitforlisten 70817 /var/tmp/spdk2.sock 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 70817 ']' 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:34.029 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:34.029 16:44:55 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:34.029 [2024-09-29 16:44:55.428327] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:34.029 [2024-09-29 16:44:55.428529] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70817 ] 00:06:34.029 [2024-09-29 16:44:55.562211] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70799 has claimed it. 00:06:34.029 [2024-09-29 16:44:55.562280] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:34.599 ERROR: process (pid: 70817) is no longer running 00:06:34.599 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (70817) - No such process 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 1 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # es=1 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 70799 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@950 -- # '[' -z 70799 ']' 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # kill -0 70799 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # uname 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70799 00:06:34.599 killing process with pid 70799 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70799' 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@969 -- # kill 70799 00:06:34.599 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@974 -- # wait 70799 00:06:34.859 00:06:34.859 real 0m2.086s 00:06:34.859 user 0m5.596s 00:06:34.859 sys 0m0.480s 00:06:34.859 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:34.859 ************************************ 00:06:34.859 END TEST locking_overlapped_coremask 00:06:34.859 ************************************ 00:06:34.859 16:44:56 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:35.119 16:44:56 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:06:35.119 16:44:56 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:35.119 16:44:56 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:35.119 16:44:56 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:35.119 ************************************ 00:06:35.119 START TEST locking_overlapped_coremask_via_rpc 00:06:35.119 ************************************ 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask_via_rpc 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=70859 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 70859 /var/tmp/spdk.sock 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70859 ']' 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:35.119 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:35.119 16:44:56 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:35.119 [2024-09-29 16:44:56.675166] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:35.119 [2024-09-29 16:44:56.675403] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70859 ] 00:06:35.379 [2024-09-29 16:44:56.818977] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:35.379 [2024-09-29 16:44:56.819048] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:35.379 [2024-09-29 16:44:56.866187] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:35.379 [2024-09-29 16:44:56.866104] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:35.379 [2024-09-29 16:44:56.866316] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=70877 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 70877 /var/tmp/spdk2.sock 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70877 ']' 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:35.948 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:35.948 16:44:57 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:35.948 [2024-09-29 16:44:57.565860] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:35.948 [2024-09-29 16:44:57.566067] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70877 ] 00:06:36.208 [2024-09-29 16:44:57.701834] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:36.208 [2024-09-29 16:44:57.701901] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:36.208 [2024-09-29 16:44:57.804651] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:06:36.208 [2024-09-29 16:44:57.804810] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 4 00:06:36.208 [2024-09-29 16:44:57.804673] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:36.778 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # local es=0 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:36.779 [2024-09-29 16:44:58.411899] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70859 has claimed it. 00:06:36.779 request: 00:06:36.779 { 00:06:36.779 "method": "framework_enable_cpumask_locks", 00:06:36.779 "req_id": 1 00:06:36.779 } 00:06:36.779 Got JSON-RPC error response 00:06:36.779 response: 00:06:36.779 { 00:06:36.779 "code": -32603, 00:06:36.779 "message": "Failed to claim CPU core: 2" 00:06:36.779 } 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # es=1 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 70859 /var/tmp/spdk.sock 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70859 ']' 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:36.779 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:36.779 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 70877 /var/tmp/spdk2.sock 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70877 ']' 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:37.039 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:37.039 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:06:37.299 00:06:37.299 real 0m2.255s 00:06:37.299 user 0m1.020s 00:06:37.299 sys 0m0.171s 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:37.299 16:44:58 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:37.299 ************************************ 00:06:37.299 END TEST locking_overlapped_coremask_via_rpc 00:06:37.299 ************************************ 00:06:37.299 16:44:58 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:06:37.299 16:44:58 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70859 ]] 00:06:37.299 16:44:58 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70859 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70859 ']' 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70859 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70859 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70859' 00:06:37.299 killing process with pid 70859 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 70859 00:06:37.299 16:44:58 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 70859 00:06:37.868 16:44:59 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70877 ]] 00:06:37.868 16:44:59 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70877 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70877 ']' 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70877 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70877 00:06:37.868 killing process with pid 70877 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70877' 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 70877 00:06:37.868 16:44:59 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 70877 00:06:38.128 16:44:59 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:38.128 16:44:59 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:06:38.128 16:44:59 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70859 ]] 00:06:38.128 16:44:59 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70859 00:06:38.128 Process with pid 70859 is not found 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70859 ']' 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70859 00:06:38.128 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (70859) - No such process 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 70859 is not found' 00:06:38.128 16:44:59 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70877 ]] 00:06:38.128 16:44:59 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70877 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70877 ']' 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70877 00:06:38.128 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (70877) - No such process 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 70877 is not found' 00:06:38.128 Process with pid 70877 is not found 00:06:38.128 16:44:59 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:06:38.128 00:06:38.128 real 0m19.263s 00:06:38.128 user 0m31.645s 00:06:38.128 sys 0m5.892s 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:38.128 16:44:59 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:38.128 ************************************ 00:06:38.128 END TEST cpu_locks 00:06:38.128 ************************************ 00:06:38.388 00:06:38.388 real 0m47.622s 00:06:38.388 user 1m29.461s 00:06:38.388 sys 0m9.777s 00:06:38.388 16:44:59 event -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:38.388 16:44:59 event -- common/autotest_common.sh@10 -- # set +x 00:06:38.388 ************************************ 00:06:38.388 END TEST event 00:06:38.388 ************************************ 00:06:38.388 16:44:59 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:38.388 16:44:59 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:38.388 16:44:59 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:38.388 16:44:59 -- common/autotest_common.sh@10 -- # set +x 00:06:38.388 ************************************ 00:06:38.388 START TEST thread 00:06:38.388 ************************************ 00:06:38.388 16:44:59 thread -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:06:38.388 * Looking for test storage... 00:06:38.388 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:06:38.388 16:45:00 thread -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:38.388 16:45:00 thread -- common/autotest_common.sh@1681 -- # lcov --version 00:06:38.388 16:45:00 thread -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:38.648 16:45:00 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:38.648 16:45:00 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:38.648 16:45:00 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:38.648 16:45:00 thread -- scripts/common.sh@336 -- # IFS=.-: 00:06:38.648 16:45:00 thread -- scripts/common.sh@336 -- # read -ra ver1 00:06:38.648 16:45:00 thread -- scripts/common.sh@337 -- # IFS=.-: 00:06:38.648 16:45:00 thread -- scripts/common.sh@337 -- # read -ra ver2 00:06:38.648 16:45:00 thread -- scripts/common.sh@338 -- # local 'op=<' 00:06:38.648 16:45:00 thread -- scripts/common.sh@340 -- # ver1_l=2 00:06:38.648 16:45:00 thread -- scripts/common.sh@341 -- # ver2_l=1 00:06:38.648 16:45:00 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:38.648 16:45:00 thread -- scripts/common.sh@344 -- # case "$op" in 00:06:38.648 16:45:00 thread -- scripts/common.sh@345 -- # : 1 00:06:38.648 16:45:00 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:38.648 16:45:00 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:38.648 16:45:00 thread -- scripts/common.sh@365 -- # decimal 1 00:06:38.648 16:45:00 thread -- scripts/common.sh@353 -- # local d=1 00:06:38.648 16:45:00 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:38.648 16:45:00 thread -- scripts/common.sh@355 -- # echo 1 00:06:38.648 16:45:00 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:06:38.648 16:45:00 thread -- scripts/common.sh@366 -- # decimal 2 00:06:38.648 16:45:00 thread -- scripts/common.sh@353 -- # local d=2 00:06:38.648 16:45:00 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:38.648 16:45:00 thread -- scripts/common.sh@355 -- # echo 2 00:06:38.648 16:45:00 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:06:38.648 16:45:00 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:38.648 16:45:00 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:38.648 16:45:00 thread -- scripts/common.sh@368 -- # return 0 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:38.648 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.648 --rc genhtml_branch_coverage=1 00:06:38.648 --rc genhtml_function_coverage=1 00:06:38.648 --rc genhtml_legend=1 00:06:38.648 --rc geninfo_all_blocks=1 00:06:38.648 --rc geninfo_unexecuted_blocks=1 00:06:38.648 00:06:38.648 ' 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:38.648 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.648 --rc genhtml_branch_coverage=1 00:06:38.648 --rc genhtml_function_coverage=1 00:06:38.648 --rc genhtml_legend=1 00:06:38.648 --rc geninfo_all_blocks=1 00:06:38.648 --rc geninfo_unexecuted_blocks=1 00:06:38.648 00:06:38.648 ' 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:38.648 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.648 --rc genhtml_branch_coverage=1 00:06:38.648 --rc genhtml_function_coverage=1 00:06:38.648 --rc genhtml_legend=1 00:06:38.648 --rc geninfo_all_blocks=1 00:06:38.648 --rc geninfo_unexecuted_blocks=1 00:06:38.648 00:06:38.648 ' 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:38.648 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:38.648 --rc genhtml_branch_coverage=1 00:06:38.648 --rc genhtml_function_coverage=1 00:06:38.648 --rc genhtml_legend=1 00:06:38.648 --rc geninfo_all_blocks=1 00:06:38.648 --rc geninfo_unexecuted_blocks=1 00:06:38.648 00:06:38.648 ' 00:06:38.648 16:45:00 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:38.648 16:45:00 thread -- common/autotest_common.sh@10 -- # set +x 00:06:38.648 ************************************ 00:06:38.648 START TEST thread_poller_perf 00:06:38.648 ************************************ 00:06:38.648 16:45:00 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:06:38.648 [2024-09-29 16:45:00.172264] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:38.648 [2024-09-29 16:45:00.172441] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71004 ] 00:06:38.648 [2024-09-29 16:45:00.317233] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:38.908 [2024-09-29 16:45:00.364996] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:38.908 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:06:39.846 ====================================== 00:06:39.846 busy:2297409470 (cyc) 00:06:39.846 total_run_count: 425000 00:06:39.846 tsc_hz: 2290000000 (cyc) 00:06:39.846 ====================================== 00:06:39.846 poller_cost: 5405 (cyc), 2360 (nsec) 00:06:39.846 00:06:39.846 real 0m1.330s 00:06:39.846 user 0m1.135s 00:06:39.846 sys 0m0.085s 00:06:39.846 16:45:01 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:39.846 16:45:01 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:39.846 ************************************ 00:06:39.846 END TEST thread_poller_perf 00:06:39.846 ************************************ 00:06:39.846 16:45:01 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:39.846 16:45:01 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:06:39.846 16:45:01 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:39.846 16:45:01 thread -- common/autotest_common.sh@10 -- # set +x 00:06:40.106 ************************************ 00:06:40.106 START TEST thread_poller_perf 00:06:40.106 ************************************ 00:06:40.107 16:45:01 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:06:40.107 [2024-09-29 16:45:01.570297] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:40.107 [2024-09-29 16:45:01.570420] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71045 ] 00:06:40.107 [2024-09-29 16:45:01.714900] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:40.107 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:06:40.107 [2024-09-29 16:45:01.760389] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:41.487 ====================================== 00:06:41.487 busy:2293384740 (cyc) 00:06:41.487 total_run_count: 5577000 00:06:41.487 tsc_hz: 2290000000 (cyc) 00:06:41.487 ====================================== 00:06:41.487 poller_cost: 411 (cyc), 179 (nsec) 00:06:41.487 00:06:41.487 real 0m1.322s 00:06:41.487 user 0m1.125s 00:06:41.487 sys 0m0.092s 00:06:41.487 ************************************ 00:06:41.487 END TEST thread_poller_perf 00:06:41.487 ************************************ 00:06:41.487 16:45:02 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:41.487 16:45:02 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:06:41.487 16:45:02 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:06:41.487 ************************************ 00:06:41.487 END TEST thread 00:06:41.487 ************************************ 00:06:41.487 00:06:41.487 real 0m3.008s 00:06:41.487 user 0m2.414s 00:06:41.487 sys 0m0.386s 00:06:41.487 16:45:02 thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:41.487 16:45:02 thread -- common/autotest_common.sh@10 -- # set +x 00:06:41.487 16:45:02 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:06:41.487 16:45:02 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:41.487 16:45:02 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:41.487 16:45:02 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:41.487 16:45:02 -- common/autotest_common.sh@10 -- # set +x 00:06:41.487 ************************************ 00:06:41.487 START TEST app_cmdline 00:06:41.487 ************************************ 00:06:41.487 16:45:02 app_cmdline -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:06:41.487 * Looking for test storage... 00:06:41.487 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:41.487 16:45:03 app_cmdline -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:41.487 16:45:03 app_cmdline -- common/autotest_common.sh@1681 -- # lcov --version 00:06:41.487 16:45:03 app_cmdline -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@345 -- # : 1 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:41.747 16:45:03 app_cmdline -- scripts/common.sh@368 -- # return 0 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:41.747 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.747 --rc genhtml_branch_coverage=1 00:06:41.747 --rc genhtml_function_coverage=1 00:06:41.747 --rc genhtml_legend=1 00:06:41.747 --rc geninfo_all_blocks=1 00:06:41.747 --rc geninfo_unexecuted_blocks=1 00:06:41.747 00:06:41.747 ' 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:41.747 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.747 --rc genhtml_branch_coverage=1 00:06:41.747 --rc genhtml_function_coverage=1 00:06:41.747 --rc genhtml_legend=1 00:06:41.747 --rc geninfo_all_blocks=1 00:06:41.747 --rc geninfo_unexecuted_blocks=1 00:06:41.747 00:06:41.747 ' 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:41.747 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.747 --rc genhtml_branch_coverage=1 00:06:41.747 --rc genhtml_function_coverage=1 00:06:41.747 --rc genhtml_legend=1 00:06:41.747 --rc geninfo_all_blocks=1 00:06:41.747 --rc geninfo_unexecuted_blocks=1 00:06:41.747 00:06:41.747 ' 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:41.747 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:41.747 --rc genhtml_branch_coverage=1 00:06:41.747 --rc genhtml_function_coverage=1 00:06:41.747 --rc genhtml_legend=1 00:06:41.747 --rc geninfo_all_blocks=1 00:06:41.747 --rc geninfo_unexecuted_blocks=1 00:06:41.747 00:06:41.747 ' 00:06:41.747 16:45:03 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:06:41.747 16:45:03 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71124 00:06:41.747 16:45:03 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:06:41.747 16:45:03 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71124 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@831 -- # '[' -z 71124 ']' 00:06:41.747 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:41.747 16:45:03 app_cmdline -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:41.748 16:45:03 app_cmdline -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:41.748 16:45:03 app_cmdline -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:41.748 16:45:03 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:41.748 [2024-09-29 16:45:03.281878] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:41.748 [2024-09-29 16:45:03.281999] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71124 ] 00:06:42.007 [2024-09-29 16:45:03.424628] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:42.007 [2024-09-29 16:45:03.471856] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:42.576 16:45:04 app_cmdline -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:42.576 16:45:04 app_cmdline -- common/autotest_common.sh@864 -- # return 0 00:06:42.576 16:45:04 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:06:42.835 { 00:06:42.835 "version": "SPDK v25.01-pre git sha1 09cc66129", 00:06:42.835 "fields": { 00:06:42.835 "major": 25, 00:06:42.835 "minor": 1, 00:06:42.835 "patch": 0, 00:06:42.835 "suffix": "-pre", 00:06:42.835 "commit": "09cc66129" 00:06:42.836 } 00:06:42.836 } 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@26 -- # sort 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:06:42.836 16:45:04 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@650 -- # local es=0 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:06:42.836 16:45:04 app_cmdline -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:43.095 request: 00:06:43.095 { 00:06:43.095 "method": "env_dpdk_get_mem_stats", 00:06:43.095 "req_id": 1 00:06:43.095 } 00:06:43.095 Got JSON-RPC error response 00:06:43.095 response: 00:06:43.095 { 00:06:43.095 "code": -32601, 00:06:43.095 "message": "Method not found" 00:06:43.095 } 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@653 -- # es=1 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:43.095 16:45:04 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71124 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@950 -- # '[' -z 71124 ']' 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@954 -- # kill -0 71124 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@955 -- # uname 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71124 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:43.095 killing process with pid 71124 00:06:43.095 16:45:04 app_cmdline -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71124' 00:06:43.096 16:45:04 app_cmdline -- common/autotest_common.sh@969 -- # kill 71124 00:06:43.096 16:45:04 app_cmdline -- common/autotest_common.sh@974 -- # wait 71124 00:06:43.354 00:06:43.354 real 0m1.983s 00:06:43.354 user 0m2.184s 00:06:43.354 sys 0m0.551s 00:06:43.354 16:45:04 app_cmdline -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:43.354 16:45:04 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:43.354 ************************************ 00:06:43.354 END TEST app_cmdline 00:06:43.354 ************************************ 00:06:43.354 16:45:05 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:43.354 16:45:05 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:43.354 16:45:05 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:43.354 16:45:05 -- common/autotest_common.sh@10 -- # set +x 00:06:43.355 ************************************ 00:06:43.355 START TEST version 00:06:43.355 ************************************ 00:06:43.355 16:45:05 version -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:06:43.614 * Looking for test storage... 00:06:43.614 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:06:43.614 16:45:05 version -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:43.614 16:45:05 version -- common/autotest_common.sh@1681 -- # lcov --version 00:06:43.614 16:45:05 version -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:43.614 16:45:05 version -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:43.614 16:45:05 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:43.614 16:45:05 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:43.614 16:45:05 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:43.614 16:45:05 version -- scripts/common.sh@336 -- # IFS=.-: 00:06:43.614 16:45:05 version -- scripts/common.sh@336 -- # read -ra ver1 00:06:43.614 16:45:05 version -- scripts/common.sh@337 -- # IFS=.-: 00:06:43.614 16:45:05 version -- scripts/common.sh@337 -- # read -ra ver2 00:06:43.614 16:45:05 version -- scripts/common.sh@338 -- # local 'op=<' 00:06:43.614 16:45:05 version -- scripts/common.sh@340 -- # ver1_l=2 00:06:43.614 16:45:05 version -- scripts/common.sh@341 -- # ver2_l=1 00:06:43.614 16:45:05 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:43.614 16:45:05 version -- scripts/common.sh@344 -- # case "$op" in 00:06:43.614 16:45:05 version -- scripts/common.sh@345 -- # : 1 00:06:43.614 16:45:05 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:43.615 16:45:05 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:43.615 16:45:05 version -- scripts/common.sh@365 -- # decimal 1 00:06:43.615 16:45:05 version -- scripts/common.sh@353 -- # local d=1 00:06:43.615 16:45:05 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:43.615 16:45:05 version -- scripts/common.sh@355 -- # echo 1 00:06:43.615 16:45:05 version -- scripts/common.sh@365 -- # ver1[v]=1 00:06:43.615 16:45:05 version -- scripts/common.sh@366 -- # decimal 2 00:06:43.615 16:45:05 version -- scripts/common.sh@353 -- # local d=2 00:06:43.615 16:45:05 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:43.615 16:45:05 version -- scripts/common.sh@355 -- # echo 2 00:06:43.615 16:45:05 version -- scripts/common.sh@366 -- # ver2[v]=2 00:06:43.615 16:45:05 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:43.615 16:45:05 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:43.615 16:45:05 version -- scripts/common.sh@368 -- # return 0 00:06:43.615 16:45:05 version -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:43.615 16:45:05 version -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:43.615 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.615 --rc genhtml_branch_coverage=1 00:06:43.615 --rc genhtml_function_coverage=1 00:06:43.615 --rc genhtml_legend=1 00:06:43.615 --rc geninfo_all_blocks=1 00:06:43.615 --rc geninfo_unexecuted_blocks=1 00:06:43.615 00:06:43.615 ' 00:06:43.615 16:45:05 version -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:43.615 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.615 --rc genhtml_branch_coverage=1 00:06:43.615 --rc genhtml_function_coverage=1 00:06:43.615 --rc genhtml_legend=1 00:06:43.615 --rc geninfo_all_blocks=1 00:06:43.615 --rc geninfo_unexecuted_blocks=1 00:06:43.615 00:06:43.615 ' 00:06:43.615 16:45:05 version -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:43.615 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.615 --rc genhtml_branch_coverage=1 00:06:43.615 --rc genhtml_function_coverage=1 00:06:43.615 --rc genhtml_legend=1 00:06:43.615 --rc geninfo_all_blocks=1 00:06:43.615 --rc geninfo_unexecuted_blocks=1 00:06:43.615 00:06:43.615 ' 00:06:43.615 16:45:05 version -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:43.615 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.615 --rc genhtml_branch_coverage=1 00:06:43.615 --rc genhtml_function_coverage=1 00:06:43.615 --rc genhtml_legend=1 00:06:43.615 --rc geninfo_all_blocks=1 00:06:43.615 --rc geninfo_unexecuted_blocks=1 00:06:43.615 00:06:43.615 ' 00:06:43.615 16:45:05 version -- app/version.sh@17 -- # get_header_version major 00:06:43.615 16:45:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # cut -f2 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:43.615 16:45:05 version -- app/version.sh@17 -- # major=25 00:06:43.615 16:45:05 version -- app/version.sh@18 -- # get_header_version minor 00:06:43.615 16:45:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # cut -f2 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:43.615 16:45:05 version -- app/version.sh@18 -- # minor=1 00:06:43.615 16:45:05 version -- app/version.sh@19 -- # get_header_version patch 00:06:43.615 16:45:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # cut -f2 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:43.615 16:45:05 version -- app/version.sh@19 -- # patch=0 00:06:43.615 16:45:05 version -- app/version.sh@20 -- # get_header_version suffix 00:06:43.615 16:45:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # cut -f2 00:06:43.615 16:45:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:43.615 16:45:05 version -- app/version.sh@20 -- # suffix=-pre 00:06:43.615 16:45:05 version -- app/version.sh@22 -- # version=25.1 00:06:43.615 16:45:05 version -- app/version.sh@25 -- # (( patch != 0 )) 00:06:43.615 16:45:05 version -- app/version.sh@28 -- # version=25.1rc0 00:06:43.615 16:45:05 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:06:43.615 16:45:05 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:06:43.875 16:45:05 version -- app/version.sh@30 -- # py_version=25.1rc0 00:06:43.875 16:45:05 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:06:43.875 ************************************ 00:06:43.875 END TEST version 00:06:43.875 ************************************ 00:06:43.875 00:06:43.875 real 0m0.312s 00:06:43.875 user 0m0.187s 00:06:43.875 sys 0m0.184s 00:06:43.875 16:45:05 version -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:43.875 16:45:05 version -- common/autotest_common.sh@10 -- # set +x 00:06:43.875 16:45:05 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:06:43.875 16:45:05 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:06:43.875 16:45:05 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:43.875 16:45:05 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:43.875 16:45:05 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:43.875 16:45:05 -- common/autotest_common.sh@10 -- # set +x 00:06:43.875 ************************************ 00:06:43.875 START TEST bdev_raid 00:06:43.875 ************************************ 00:06:43.875 16:45:05 bdev_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:06:43.875 * Looking for test storage... 00:06:43.875 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:06:43.875 16:45:05 bdev_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:43.875 16:45:05 bdev_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:06:43.875 16:45:05 bdev_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:44.134 16:45:05 bdev_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@345 -- # : 1 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:44.134 16:45:05 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:44.135 16:45:05 bdev_raid -- scripts/common.sh@368 -- # return 0 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:44.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:44.135 --rc genhtml_branch_coverage=1 00:06:44.135 --rc genhtml_function_coverage=1 00:06:44.135 --rc genhtml_legend=1 00:06:44.135 --rc geninfo_all_blocks=1 00:06:44.135 --rc geninfo_unexecuted_blocks=1 00:06:44.135 00:06:44.135 ' 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:44.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:44.135 --rc genhtml_branch_coverage=1 00:06:44.135 --rc genhtml_function_coverage=1 00:06:44.135 --rc genhtml_legend=1 00:06:44.135 --rc geninfo_all_blocks=1 00:06:44.135 --rc geninfo_unexecuted_blocks=1 00:06:44.135 00:06:44.135 ' 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:44.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:44.135 --rc genhtml_branch_coverage=1 00:06:44.135 --rc genhtml_function_coverage=1 00:06:44.135 --rc genhtml_legend=1 00:06:44.135 --rc geninfo_all_blocks=1 00:06:44.135 --rc geninfo_unexecuted_blocks=1 00:06:44.135 00:06:44.135 ' 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:44.135 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:44.135 --rc genhtml_branch_coverage=1 00:06:44.135 --rc genhtml_function_coverage=1 00:06:44.135 --rc genhtml_legend=1 00:06:44.135 --rc geninfo_all_blocks=1 00:06:44.135 --rc geninfo_unexecuted_blocks=1 00:06:44.135 00:06:44.135 ' 00:06:44.135 16:45:05 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:44.135 16:45:05 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:06:44.135 16:45:05 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:06:44.135 16:45:05 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:06:44.135 16:45:05 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:06:44.135 16:45:05 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:06:44.135 16:45:05 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:44.135 16:45:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:44.135 ************************************ 00:06:44.135 START TEST raid1_resize_data_offset_test 00:06:44.135 ************************************ 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1125 -- # raid_resize_data_offset_test 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71290 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:44.135 Process raid pid: 71290 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71290' 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71290 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@831 -- # '[' -z 71290 ']' 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:44.135 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:44.135 16:45:05 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.135 [2024-09-29 16:45:05.718200] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:44.135 [2024-09-29 16:45:05.718332] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:44.395 [2024-09-29 16:45:05.863575] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:44.395 [2024-09-29 16:45:05.909118] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:44.395 [2024-09-29 16:45:05.950830] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:44.395 [2024-09-29 16:45:05.950882] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # return 0 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.963 malloc0 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.963 malloc1 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.963 null0 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:44.963 [2024-09-29 16:45:06.618410] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:06:44.963 [2024-09-29 16:45:06.620258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:06:44.963 [2024-09-29 16:45:06.620327] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:06:44.963 [2024-09-29 16:45:06.620470] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:44.963 [2024-09-29 16:45:06.620481] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:06:44.963 [2024-09-29 16:45:06.620793] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:44.963 [2024-09-29 16:45:06.620973] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:44.963 [2024-09-29 16:45:06.621004] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:44.963 [2024-09-29 16:45:06.621137] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:44.963 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.222 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:45.222 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:06:45.222 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:06:45.222 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:45.222 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.222 [2024-09-29 16:45:06.674293] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:06:45.222 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.223 malloc2 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.223 [2024-09-29 16:45:06.806905] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:06:45.223 [2024-09-29 16:45:06.811818] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:45.223 [2024-09-29 16:45:06.813964] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71290 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@950 -- # '[' -z 71290 ']' 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # kill -0 71290 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # uname 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:45.223 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71290 00:06:45.482 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:45.482 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:45.482 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71290' 00:06:45.482 killing process with pid 71290 00:06:45.482 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@969 -- # kill 71290 00:06:45.482 16:45:06 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@974 -- # wait 71290 00:06:45.482 [2024-09-29 16:45:06.903063] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:45.482 [2024-09-29 16:45:06.904754] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:06:45.482 [2024-09-29 16:45:06.904809] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:45.482 [2024-09-29 16:45:06.904827] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:06:45.482 [2024-09-29 16:45:06.910192] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:45.482 [2024-09-29 16:45:06.910458] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:45.482 [2024-09-29 16:45:06.910477] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:45.482 [2024-09-29 16:45:07.120485] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:45.741 16:45:07 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:06:45.741 00:06:45.741 real 0m1.721s 00:06:45.741 user 0m1.716s 00:06:45.741 sys 0m0.435s 00:06:45.741 16:45:07 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:45.741 16:45:07 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:06:45.741 ************************************ 00:06:45.741 END TEST raid1_resize_data_offset_test 00:06:45.741 ************************************ 00:06:46.001 16:45:07 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:06:46.001 16:45:07 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:46.001 16:45:07 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:46.001 16:45:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:46.001 ************************************ 00:06:46.001 START TEST raid0_resize_superblock_test 00:06:46.001 ************************************ 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 0 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71340 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:46.001 Process raid pid: 71340 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71340' 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71340 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 71340 ']' 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:46.001 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:46.001 16:45:07 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.001 [2024-09-29 16:45:07.511193] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:46.001 [2024-09-29 16:45:07.511307] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:46.001 [2024-09-29 16:45:07.655893] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:46.260 [2024-09-29 16:45:07.702343] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:46.260 [2024-09-29 16:45:07.744379] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:46.260 [2024-09-29 16:45:07.744418] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.828 malloc0 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:46.828 [2024-09-29 16:45:08.462518] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:46.828 [2024-09-29 16:45:08.462572] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:46.828 [2024-09-29 16:45:08.462594] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:46.828 [2024-09-29 16:45:08.462611] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:46.828 [2024-09-29 16:45:08.464732] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:46.828 [2024-09-29 16:45:08.464767] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:46.828 pt0 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.828 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.087 2e3ee85d-e18c-40ed-8b30-ae9be0929bbe 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.087 58c97dfa-86e3-4a32-99ec-c55a55dd7ec9 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.087 51352280-cbce-4c25-8bb0-39c0296257b6 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.087 [2024-09-29 16:45:08.598426] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 58c97dfa-86e3-4a32-99ec-c55a55dd7ec9 is claimed 00:06:47.087 [2024-09-29 16:45:08.598505] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 51352280-cbce-4c25-8bb0-39c0296257b6 is claimed 00:06:47.087 [2024-09-29 16:45:08.598616] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:47.087 [2024-09-29 16:45:08.598629] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:06:47.087 [2024-09-29 16:45:08.598931] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:47.087 [2024-09-29 16:45:08.599093] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:47.087 [2024-09-29 16:45:08.599118] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:47.087 [2024-09-29 16:45:08.599300] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:06:47.087 [2024-09-29 16:45:08.710435] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:47.087 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.088 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.088 [2024-09-29 16:45:08.758334] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:47.088 [2024-09-29 16:45:08.758361] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '58c97dfa-86e3-4a32-99ec-c55a55dd7ec9' was resized: old size 131072, new size 204800 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.347 [2024-09-29 16:45:08.770207] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:47.347 [2024-09-29 16:45:08.770233] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '51352280-cbce-4c25-8bb0-39c0296257b6' was resized: old size 131072, new size 204800 00:06:47.347 [2024-09-29 16:45:08.770251] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:06:47.347 [2024-09-29 16:45:08.866130] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.347 [2024-09-29 16:45:08.901891] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:47.347 [2024-09-29 16:45:08.901945] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:47.347 [2024-09-29 16:45:08.901958] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:47.347 [2024-09-29 16:45:08.901968] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:47.347 [2024-09-29 16:45:08.902062] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:47.347 [2024-09-29 16:45:08.902091] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:47.347 [2024-09-29 16:45:08.902102] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.347 [2024-09-29 16:45:08.913844] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:47.347 [2024-09-29 16:45:08.913887] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:47.347 [2024-09-29 16:45:08.913903] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:06:47.347 [2024-09-29 16:45:08.913914] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:47.347 [2024-09-29 16:45:08.916078] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:47.347 [2024-09-29 16:45:08.916112] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:47.347 [2024-09-29 16:45:08.917484] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 58c97dfa-86e3-4a32-99ec-c55a55dd7ec9 00:06:47.347 [2024-09-29 16:45:08.917549] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 58c97dfa-86e3-4a32-99ec-c55a55dd7ec9 is claimed 00:06:47.347 [2024-09-29 16:45:08.917641] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 51352280-cbce-4c25-8bb0-39c0296257b6 00:06:47.347 [2024-09-29 16:45:08.917685] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 51352280-cbce-4c25-8bb0-39c0296257b6 is claimed 00:06:47.347 [2024-09-29 16:45:08.917794] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 51352280-cbce-4c25-8bb0-39c0296257b6 (2) smaller than existing raid bdev Raid (3) 00:06:47.347 [2024-09-29 16:45:08.917825] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 58c97dfa-86e3-4a32-99ec-c55a55dd7ec9: File exists 00:06:47.347 [2024-09-29 16:45:08.917859] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:06:47.347 [2024-09-29 16:45:08.917869] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:06:47.347 [2024-09-29 16:45:08.918115] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:06:47.347 [2024-09-29 16:45:08.918274] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:06:47.347 [2024-09-29 16:45:08.918295] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:06:47.347 [2024-09-29 16:45:08.918453] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:47.347 pt0 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:47.347 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.348 [2024-09-29 16:45:08.942049] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71340 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 71340 ']' 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 71340 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:47.348 16:45:08 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71340 00:06:47.606 16:45:09 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:47.606 16:45:09 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:47.606 killing process with pid 71340 00:06:47.606 16:45:09 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71340' 00:06:47.606 16:45:09 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 71340 00:06:47.606 [2024-09-29 16:45:09.022295] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:47.606 [2024-09-29 16:45:09.022361] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:47.606 [2024-09-29 16:45:09.022400] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:47.606 [2024-09-29 16:45:09.022410] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:06:47.606 16:45:09 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 71340 00:06:47.606 [2024-09-29 16:45:09.181081] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:47.864 16:45:09 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:47.864 00:06:47.864 real 0m1.987s 00:06:47.864 user 0m2.255s 00:06:47.864 sys 0m0.452s 00:06:47.864 ************************************ 00:06:47.864 END TEST raid0_resize_superblock_test 00:06:47.864 ************************************ 00:06:47.864 16:45:09 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:47.864 16:45:09 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:47.864 16:45:09 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:06:47.864 16:45:09 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:47.864 16:45:09 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:47.864 16:45:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:47.864 ************************************ 00:06:47.864 START TEST raid1_resize_superblock_test 00:06:47.864 ************************************ 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 1 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71411 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:47.864 Process raid pid: 71411 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71411' 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71411 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 71411 ']' 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:47.864 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:47.864 16:45:09 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.123 [2024-09-29 16:45:09.568011] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:48.123 [2024-09-29 16:45:09.568122] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:48.123 [2024-09-29 16:45:09.712113] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:48.123 [2024-09-29 16:45:09.757662] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:48.387 [2024-09-29 16:45:09.799595] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:48.387 [2024-09-29 16:45:09.799637] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.972 malloc0 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.972 [2024-09-29 16:45:10.505767] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:48.972 [2024-09-29 16:45:10.505839] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:48.972 [2024-09-29 16:45:10.505859] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:06:48.972 [2024-09-29 16:45:10.505869] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:48.972 [2024-09-29 16:45:10.508007] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:48.972 [2024-09-29 16:45:10.508045] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:48.972 pt0 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.972 7435f8c3-59c5-4bdc-a1a2-48334f20e38b 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.972 57a97f80-8c13-44b0-a2b9-7eb98ce3b38f 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.972 20cd4d32-a242-4598-aa73-89508edeaf42 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:48.972 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:48.972 [2024-09-29 16:45:10.642038] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 57a97f80-8c13-44b0-a2b9-7eb98ce3b38f is claimed 00:06:48.972 [2024-09-29 16:45:10.642129] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 20cd4d32-a242-4598-aa73-89508edeaf42 is claimed 00:06:48.972 [2024-09-29 16:45:10.642255] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:48.972 [2024-09-29 16:45:10.642269] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:06:48.972 [2024-09-29 16:45:10.642570] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:48.972 [2024-09-29 16:45:10.642776] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:48.972 [2024-09-29 16:45:10.642797] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:48.972 [2024-09-29 16:45:10.642967] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 [2024-09-29 16:45:10.738068] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 [2024-09-29 16:45:10.781938] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:49.231 [2024-09-29 16:45:10.781967] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '57a97f80-8c13-44b0-a2b9-7eb98ce3b38f' was resized: old size 131072, new size 204800 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 [2024-09-29 16:45:10.789868] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:49.231 [2024-09-29 16:45:10.789894] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '20cd4d32-a242-4598-aa73-89508edeaf42' was resized: old size 131072, new size 204800 00:06:49.231 [2024-09-29 16:45:10.789913] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:06:49.231 [2024-09-29 16:45:10.877810] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:49.231 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.492 [2024-09-29 16:45:10.929542] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:06:49.492 [2024-09-29 16:45:10.929602] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:06:49.492 [2024-09-29 16:45:10.929635] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:06:49.492 [2024-09-29 16:45:10.929802] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:06:49.492 [2024-09-29 16:45:10.929933] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:49.492 [2024-09-29 16:45:10.929986] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:49.492 [2024-09-29 16:45:10.929997] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.492 [2024-09-29 16:45:10.941479] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:06:49.492 [2024-09-29 16:45:10.941536] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:06:49.492 [2024-09-29 16:45:10.941553] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:06:49.492 [2024-09-29 16:45:10.941563] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:06:49.492 [2024-09-29 16:45:10.943608] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:06:49.492 [2024-09-29 16:45:10.943641] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:06:49.492 [2024-09-29 16:45:10.944985] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 57a97f80-8c13-44b0-a2b9-7eb98ce3b38f 00:06:49.492 [2024-09-29 16:45:10.945038] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 57a97f80-8c13-44b0-a2b9-7eb98ce3b38f is claimed 00:06:49.492 [2024-09-29 16:45:10.945121] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 20cd4d32-a242-4598-aa73-89508edeaf42 00:06:49.492 [2024-09-29 16:45:10.945147] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 20cd4d32-a242-4598-aa73-89508edeaf42 is claimed 00:06:49.492 [2024-09-29 16:45:10.945237] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 20cd4d32-a242-4598-aa73-89508edeaf42 (2) smaller than existing raid bdev Raid (3) 00:06:49.492 [2024-09-29 16:45:10.945272] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 57a97f80-8c13-44b0-a2b9-7eb98ce3b38f: File exists 00:06:49.492 [2024-09-29 16:45:10.945303] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:06:49.492 [2024-09-29 16:45:10.945311] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:06:49.492 [2024-09-29 16:45:10.945570] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:06:49.492 [2024-09-29 16:45:10.945740] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:06:49.492 [2024-09-29 16:45:10.945758] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:06:49.492 [2024-09-29 16:45:10.945908] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:49.492 pt0 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:49.492 [2024-09-29 16:45:10.970044] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.492 16:45:10 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71411 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 71411 ']' 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 71411 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71411 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:49.492 killing process with pid 71411 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71411' 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 71411 00:06:49.492 [2024-09-29 16:45:11.050359] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:49.492 [2024-09-29 16:45:11.050427] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:49.492 [2024-09-29 16:45:11.050497] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:49.492 [2024-09-29 16:45:11.050508] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:06:49.492 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 71411 00:06:49.753 [2024-09-29 16:45:11.210008] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:50.013 16:45:11 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:06:50.013 00:06:50.013 real 0m1.954s 00:06:50.013 user 0m2.213s 00:06:50.013 sys 0m0.476s 00:06:50.013 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:50.013 16:45:11 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:06:50.013 ************************************ 00:06:50.013 END TEST raid1_resize_superblock_test 00:06:50.013 ************************************ 00:06:50.013 16:45:11 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:06:50.013 16:45:11 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:06:50.013 16:45:11 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:06:50.013 16:45:11 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:06:50.013 16:45:11 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:06:50.013 16:45:11 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:06:50.013 16:45:11 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:50.013 16:45:11 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:50.013 16:45:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:50.013 ************************************ 00:06:50.013 START TEST raid_function_test_raid0 00:06:50.013 ************************************ 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1125 -- # raid_function_test raid0 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71487 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:50.013 Process raid pid: 71487 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71487' 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71487 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@831 -- # '[' -z 71487 ']' 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:50.013 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:50.013 16:45:11 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:50.013 [2024-09-29 16:45:11.619511] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:50.013 [2024-09-29 16:45:11.619634] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:50.272 [2024-09-29 16:45:11.766600] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:50.272 [2024-09-29 16:45:11.813161] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:50.273 [2024-09-29 16:45:11.854963] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:50.273 [2024-09-29 16:45:11.855002] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # return 0 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:50.840 Base_1 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:50.840 Base_2 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:50.840 [2024-09-29 16:45:12.495989] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:50.840 [2024-09-29 16:45:12.500194] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:50.840 [2024-09-29 16:45:12.500342] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:50.840 [2024-09-29 16:45:12.500370] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:50.840 [2024-09-29 16:45:12.501094] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:50.840 [2024-09-29 16:45:12.501414] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:50.840 [2024-09-29 16:45:12.501453] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:06:50.840 [2024-09-29 16:45:12.501806] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:50.840 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:51.100 [2024-09-29 16:45:12.736085] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:06:51.100 /dev/nbd0 00:06:51.100 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@869 -- # local i 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # break 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:51.360 1+0 records in 00:06:51.360 1+0 records out 00:06:51.360 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000260471 s, 15.7 MB/s 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # size=4096 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # return 0 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:51.360 { 00:06:51.360 "nbd_device": "/dev/nbd0", 00:06:51.360 "bdev_name": "raid" 00:06:51.360 } 00:06:51.360 ]' 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:51.360 { 00:06:51.360 "nbd_device": "/dev/nbd0", 00:06:51.360 "bdev_name": "raid" 00:06:51.360 } 00:06:51.360 ]' 00:06:51.360 16:45:12 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:51.620 4096+0 records in 00:06:51.620 4096+0 records out 00:06:51.620 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0370596 s, 56.6 MB/s 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:51.620 4096+0 records in 00:06:51.620 4096+0 records out 00:06:51.620 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.181946 s, 11.5 MB/s 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:51.620 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:51.881 128+0 records in 00:06:51.881 128+0 records out 00:06:51.881 65536 bytes (66 kB, 64 KiB) copied, 0.00119613 s, 54.8 MB/s 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:51.881 2035+0 records in 00:06:51.881 2035+0 records out 00:06:51.881 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0133816 s, 77.9 MB/s 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:51.881 456+0 records in 00:06:51.881 456+0 records out 00:06:51.881 233472 bytes (233 kB, 228 KiB) copied, 0.0030372 s, 76.9 MB/s 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:51.881 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:52.141 [2024-09-29 16:45:13.599068] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:52.141 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71487 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@950 -- # '[' -z 71487 ']' 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # kill -0 71487 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # uname 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71487 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:52.401 killing process with pid 71487 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71487' 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@969 -- # kill 71487 00:06:52.401 [2024-09-29 16:45:13.910529] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:52.401 [2024-09-29 16:45:13.910642] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:52.401 [2024-09-29 16:45:13.910701] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:52.401 16:45:13 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@974 -- # wait 71487 00:06:52.401 [2024-09-29 16:45:13.910713] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:06:52.401 [2024-09-29 16:45:13.933150] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:52.661 16:45:14 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:06:52.661 00:06:52.661 real 0m2.631s 00:06:52.661 user 0m3.211s 00:06:52.661 sys 0m0.914s 00:06:52.661 16:45:14 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:52.661 16:45:14 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:06:52.661 ************************************ 00:06:52.661 END TEST raid_function_test_raid0 00:06:52.661 ************************************ 00:06:52.661 16:45:14 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:06:52.661 16:45:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:52.661 16:45:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:52.661 16:45:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:52.661 ************************************ 00:06:52.661 START TEST raid_function_test_concat 00:06:52.661 ************************************ 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1125 -- # raid_function_test concat 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=71609 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71609' 00:06:52.661 Process raid pid: 71609 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 71609 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@831 -- # '[' -z 71609 ']' 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:52.661 16:45:14 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:52.662 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:52.662 16:45:14 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:52.662 16:45:14 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:52.662 16:45:14 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:52.662 [2024-09-29 16:45:14.316253] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:52.662 [2024-09-29 16:45:14.316365] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:52.922 [2024-09-29 16:45:14.462176] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:52.922 [2024-09-29 16:45:14.507505] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:52.922 [2024-09-29 16:45:14.549959] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:52.922 [2024-09-29 16:45:14.549995] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # return 0 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:53.492 Base_1 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:53.492 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:53.751 Base_2 00:06:53.751 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:53.751 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:06:53.751 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:53.751 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:53.751 [2024-09-29 16:45:15.184362] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:53.751 [2024-09-29 16:45:15.188257] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:53.751 [2024-09-29 16:45:15.188372] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:53.751 [2024-09-29 16:45:15.188392] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:53.751 [2024-09-29 16:45:15.188825] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:06:53.752 [2024-09-29 16:45:15.189063] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:53.752 [2024-09-29 16:45:15.189098] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:06:53.752 [2024-09-29 16:45:15.189371] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:53.752 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:06:53.752 [2024-09-29 16:45:15.400139] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:06:53.752 /dev/nbd0 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@869 -- # local i 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # break 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:06:54.011 1+0 records in 00:06:54.011 1+0 records out 00:06:54.011 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000461941 s, 8.9 MB/s 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # size=4096 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:54.011 16:45:15 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # return 0 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:54.012 { 00:06:54.012 "nbd_device": "/dev/nbd0", 00:06:54.012 "bdev_name": "raid" 00:06:54.012 } 00:06:54.012 ]' 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:54.012 { 00:06:54.012 "nbd_device": "/dev/nbd0", 00:06:54.012 "bdev_name": "raid" 00:06:54.012 } 00:06:54.012 ]' 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:06:54.012 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:06:54.271 4096+0 records in 00:06:54.271 4096+0 records out 00:06:54.271 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0324974 s, 64.5 MB/s 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:06:54.271 4096+0 records in 00:06:54.271 4096+0 records out 00:06:54.271 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.177061 s, 11.8 MB/s 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:06:54.271 128+0 records in 00:06:54.271 128+0 records out 00:06:54.271 65536 bytes (66 kB, 64 KiB) copied, 0.00110912 s, 59.1 MB/s 00:06:54.271 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:06:54.531 2035+0 records in 00:06:54.531 2035+0 records out 00:06:54.531 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0133928 s, 77.8 MB/s 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:06:54.531 16:45:15 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:06:54.531 456+0 records in 00:06:54.531 456+0 records out 00:06:54.531 233472 bytes (233 kB, 228 KiB) copied, 0.00350809 s, 66.6 MB/s 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:54.531 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:54.790 [2024-09-29 16:45:16.222228] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:54.790 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 71609 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@950 -- # '[' -z 71609 ']' 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # kill -0 71609 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # uname 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71609 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71609' 00:06:55.051 killing process with pid 71609 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@969 -- # kill 71609 00:06:55.051 [2024-09-29 16:45:16.533504] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:55.051 [2024-09-29 16:45:16.533639] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:55.051 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@974 -- # wait 71609 00:06:55.051 [2024-09-29 16:45:16.533697] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:55.051 [2024-09-29 16:45:16.533712] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:06:55.051 [2024-09-29 16:45:16.556309] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:55.312 16:45:16 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:06:55.312 00:06:55.312 real 0m2.558s 00:06:55.312 user 0m3.119s 00:06:55.312 sys 0m0.873s 00:06:55.312 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:55.312 16:45:16 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:06:55.312 ************************************ 00:06:55.312 END TEST raid_function_test_concat 00:06:55.312 ************************************ 00:06:55.312 16:45:16 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:06:55.312 16:45:16 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:55.312 16:45:16 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:55.312 16:45:16 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:55.312 ************************************ 00:06:55.312 START TEST raid0_resize_test 00:06:55.312 ************************************ 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 0 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71721 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71721' 00:06:55.312 Process raid pid: 71721 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71721 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- common/autotest_common.sh@831 -- # '[' -z 71721 ']' 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:55.312 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:55.312 16:45:16 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:55.312 [2024-09-29 16:45:16.940486] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:55.312 [2024-09-29 16:45:16.940612] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:55.572 [2024-09-29 16:45:17.086364] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:55.572 [2024-09-29 16:45:17.131760] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:55.572 [2024-09-29 16:45:17.173192] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:55.572 [2024-09-29 16:45:17.173232] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # return 0 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.143 Base_1 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.143 Base_2 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.143 [2024-09-29 16:45:17.782171] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:56.143 [2024-09-29 16:45:17.783937] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:56.143 [2024-09-29 16:45:17.783995] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:56.143 [2024-09-29 16:45:17.784006] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:06:56.143 [2024-09-29 16:45:17.784257] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:56.143 [2024-09-29 16:45:17.784359] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:56.143 [2024-09-29 16:45:17.784372] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:56.143 [2024-09-29 16:45:17.784514] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.143 [2024-09-29 16:45:17.794134] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:56.143 [2024-09-29 16:45:17.794163] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:56.143 true 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:56.143 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.143 [2024-09-29 16:45:17.810265] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.403 [2024-09-29 16:45:17.854016] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:56.403 [2024-09-29 16:45:17.854044] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:56.403 [2024-09-29 16:45:17.854067] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:06:56.403 true 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.403 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:56.404 [2024-09-29 16:45:17.866148] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71721 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@950 -- # '[' -z 71721 ']' 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # kill -0 71721 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # uname 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71721 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:56.404 killing process with pid 71721 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71721' 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@969 -- # kill 71721 00:06:56.404 [2024-09-29 16:45:17.938022] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:56.404 [2024-09-29 16:45:17.938116] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:56.404 [2024-09-29 16:45:17.938173] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:56.404 [2024-09-29 16:45:17.938193] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:56.404 16:45:17 bdev_raid.raid0_resize_test -- common/autotest_common.sh@974 -- # wait 71721 00:06:56.404 [2024-09-29 16:45:17.939707] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:56.664 16:45:18 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:56.664 00:06:56.664 real 0m1.318s 00:06:56.664 user 0m1.468s 00:06:56.664 sys 0m0.283s 00:06:56.664 16:45:18 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:56.664 16:45:18 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.664 ************************************ 00:06:56.664 END TEST raid0_resize_test 00:06:56.664 ************************************ 00:06:56.664 16:45:18 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:06:56.664 16:45:18 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:06:56.664 16:45:18 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:56.664 16:45:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:56.664 ************************************ 00:06:56.664 START TEST raid1_resize_test 00:06:56.664 ************************************ 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 1 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71766 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71766' 00:06:56.664 Process raid pid: 71766 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71766 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- common/autotest_common.sh@831 -- # '[' -z 71766 ']' 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:56.664 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:56.664 16:45:18 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:56.664 [2024-09-29 16:45:18.328659] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:56.664 [2024-09-29 16:45:18.328793] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:56.924 [2024-09-29 16:45:18.457120] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:56.924 [2024-09-29 16:45:18.502564] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:56.924 [2024-09-29 16:45:18.544515] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:56.924 [2024-09-29 16:45:18.544555] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # return 0 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.863 Base_1 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.863 Base_2 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.863 [2024-09-29 16:45:19.197461] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:06:57.863 [2024-09-29 16:45:19.199275] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:06:57.863 [2024-09-29 16:45:19.199336] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:06:57.863 [2024-09-29 16:45:19.199347] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:06:57.863 [2024-09-29 16:45:19.199606] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:06:57.863 [2024-09-29 16:45:19.199734] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:06:57.863 [2024-09-29 16:45:19.199747] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:06:57.863 [2024-09-29 16:45:19.199876] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.863 [2024-09-29 16:45:19.209405] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:57.863 [2024-09-29 16:45:19.209437] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:06:57.863 true 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.863 [2024-09-29 16:45:19.225536] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.863 [2024-09-29 16:45:19.253322] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:06:57.863 [2024-09-29 16:45:19.253347] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:06:57.863 [2024-09-29 16:45:19.253365] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:06:57.863 true 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:57.863 [2024-09-29 16:45:19.269464] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:06:57.863 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71766 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@950 -- # '[' -z 71766 ']' 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # kill -0 71766 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # uname 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71766 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:57.864 killing process with pid 71766 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71766' 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@969 -- # kill 71766 00:06:57.864 [2024-09-29 16:45:19.352299] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:06:57.864 [2024-09-29 16:45:19.352396] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:06:57.864 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@974 -- # wait 71766 00:06:57.864 [2024-09-29 16:45:19.352827] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:06:57.864 [2024-09-29 16:45:19.352853] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:06:57.864 [2024-09-29 16:45:19.353999] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:06:58.124 16:45:19 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:06:58.124 00:06:58.124 real 0m1.344s 00:06:58.124 user 0m1.524s 00:06:58.124 sys 0m0.279s 00:06:58.124 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:58.124 16:45:19 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.124 ************************************ 00:06:58.124 END TEST raid1_resize_test 00:06:58.124 ************************************ 00:06:58.124 16:45:19 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:06:58.124 16:45:19 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:06:58.124 16:45:19 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:06:58.124 16:45:19 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:06:58.124 16:45:19 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:58.124 16:45:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:06:58.124 ************************************ 00:06:58.124 START TEST raid_state_function_test 00:06:58.124 ************************************ 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 false 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=71817 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71817' 00:06:58.124 Process raid pid: 71817 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 71817 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 71817 ']' 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:58.124 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:58.124 16:45:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.124 [2024-09-29 16:45:19.756101] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:06:58.124 [2024-09-29 16:45:19.756206] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:58.385 [2024-09-29 16:45:19.902942] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:58.385 [2024-09-29 16:45:19.949382] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:58.385 [2024-09-29 16:45:19.991352] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:58.385 [2024-09-29 16:45:19.991391] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.955 [2024-09-29 16:45:20.580554] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:58.955 [2024-09-29 16:45:20.580602] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:58.955 [2024-09-29 16:45:20.580626] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:58.955 [2024-09-29 16:45:20.580638] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:58.955 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.214 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:59.214 "name": "Existed_Raid", 00:06:59.214 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:59.214 "strip_size_kb": 64, 00:06:59.214 "state": "configuring", 00:06:59.214 "raid_level": "raid0", 00:06:59.214 "superblock": false, 00:06:59.214 "num_base_bdevs": 2, 00:06:59.214 "num_base_bdevs_discovered": 0, 00:06:59.214 "num_base_bdevs_operational": 2, 00:06:59.214 "base_bdevs_list": [ 00:06:59.214 { 00:06:59.214 "name": "BaseBdev1", 00:06:59.214 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:59.214 "is_configured": false, 00:06:59.214 "data_offset": 0, 00:06:59.214 "data_size": 0 00:06:59.214 }, 00:06:59.214 { 00:06:59.214 "name": "BaseBdev2", 00:06:59.214 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:59.214 "is_configured": false, 00:06:59.214 "data_offset": 0, 00:06:59.214 "data_size": 0 00:06:59.214 } 00:06:59.214 ] 00:06:59.214 }' 00:06:59.214 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:59.214 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.475 16:45:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:06:59.475 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.475 16:45:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.475 [2024-09-29 16:45:21.003750] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:06:59.475 [2024-09-29 16:45:21.003798] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.475 [2024-09-29 16:45:21.011747] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:06:59.475 [2024-09-29 16:45:21.011781] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:06:59.475 [2024-09-29 16:45:21.011796] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:06:59.475 [2024-09-29 16:45:21.011806] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.475 [2024-09-29 16:45:21.028536] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:06:59.475 BaseBdev1 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.475 [ 00:06:59.475 { 00:06:59.475 "name": "BaseBdev1", 00:06:59.475 "aliases": [ 00:06:59.475 "718a3e8e-6a19-44e2-bbaa-9a79ef2f6ac7" 00:06:59.475 ], 00:06:59.475 "product_name": "Malloc disk", 00:06:59.475 "block_size": 512, 00:06:59.475 "num_blocks": 65536, 00:06:59.475 "uuid": "718a3e8e-6a19-44e2-bbaa-9a79ef2f6ac7", 00:06:59.475 "assigned_rate_limits": { 00:06:59.475 "rw_ios_per_sec": 0, 00:06:59.475 "rw_mbytes_per_sec": 0, 00:06:59.475 "r_mbytes_per_sec": 0, 00:06:59.475 "w_mbytes_per_sec": 0 00:06:59.475 }, 00:06:59.475 "claimed": true, 00:06:59.475 "claim_type": "exclusive_write", 00:06:59.475 "zoned": false, 00:06:59.475 "supported_io_types": { 00:06:59.475 "read": true, 00:06:59.475 "write": true, 00:06:59.475 "unmap": true, 00:06:59.475 "flush": true, 00:06:59.475 "reset": true, 00:06:59.475 "nvme_admin": false, 00:06:59.475 "nvme_io": false, 00:06:59.475 "nvme_io_md": false, 00:06:59.475 "write_zeroes": true, 00:06:59.475 "zcopy": true, 00:06:59.475 "get_zone_info": false, 00:06:59.475 "zone_management": false, 00:06:59.475 "zone_append": false, 00:06:59.475 "compare": false, 00:06:59.475 "compare_and_write": false, 00:06:59.475 "abort": true, 00:06:59.475 "seek_hole": false, 00:06:59.475 "seek_data": false, 00:06:59.475 "copy": true, 00:06:59.475 "nvme_iov_md": false 00:06:59.475 }, 00:06:59.475 "memory_domains": [ 00:06:59.475 { 00:06:59.475 "dma_device_id": "system", 00:06:59.475 "dma_device_type": 1 00:06:59.475 }, 00:06:59.475 { 00:06:59.475 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:59.475 "dma_device_type": 2 00:06:59.475 } 00:06:59.475 ], 00:06:59.475 "driver_specific": {} 00:06:59.475 } 00:06:59.475 ] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:06:59.475 "name": "Existed_Raid", 00:06:59.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:59.475 "strip_size_kb": 64, 00:06:59.475 "state": "configuring", 00:06:59.475 "raid_level": "raid0", 00:06:59.475 "superblock": false, 00:06:59.475 "num_base_bdevs": 2, 00:06:59.475 "num_base_bdevs_discovered": 1, 00:06:59.475 "num_base_bdevs_operational": 2, 00:06:59.475 "base_bdevs_list": [ 00:06:59.475 { 00:06:59.475 "name": "BaseBdev1", 00:06:59.475 "uuid": "718a3e8e-6a19-44e2-bbaa-9a79ef2f6ac7", 00:06:59.475 "is_configured": true, 00:06:59.475 "data_offset": 0, 00:06:59.475 "data_size": 65536 00:06:59.475 }, 00:06:59.475 { 00:06:59.475 "name": "BaseBdev2", 00:06:59.475 "uuid": "00000000-0000-0000-0000-000000000000", 00:06:59.475 "is_configured": false, 00:06:59.475 "data_offset": 0, 00:06:59.475 "data_size": 0 00:06:59.475 } 00:06:59.475 ] 00:06:59.475 }' 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:06:59.475 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.045 [2024-09-29 16:45:21.507746] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:00.045 [2024-09-29 16:45:21.507793] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.045 [2024-09-29 16:45:21.519766] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:00.045 [2024-09-29 16:45:21.521608] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:00.045 [2024-09-29 16:45:21.521642] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.045 "name": "Existed_Raid", 00:07:00.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.045 "strip_size_kb": 64, 00:07:00.045 "state": "configuring", 00:07:00.045 "raid_level": "raid0", 00:07:00.045 "superblock": false, 00:07:00.045 "num_base_bdevs": 2, 00:07:00.045 "num_base_bdevs_discovered": 1, 00:07:00.045 "num_base_bdevs_operational": 2, 00:07:00.045 "base_bdevs_list": [ 00:07:00.045 { 00:07:00.045 "name": "BaseBdev1", 00:07:00.045 "uuid": "718a3e8e-6a19-44e2-bbaa-9a79ef2f6ac7", 00:07:00.045 "is_configured": true, 00:07:00.045 "data_offset": 0, 00:07:00.045 "data_size": 65536 00:07:00.045 }, 00:07:00.045 { 00:07:00.045 "name": "BaseBdev2", 00:07:00.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:00.045 "is_configured": false, 00:07:00.045 "data_offset": 0, 00:07:00.045 "data_size": 0 00:07:00.045 } 00:07:00.045 ] 00:07:00.045 }' 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.045 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.305 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:00.305 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.305 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.305 [2024-09-29 16:45:21.968365] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:00.305 [2024-09-29 16:45:21.968417] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:00.305 [2024-09-29 16:45:21.968430] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:00.305 [2024-09-29 16:45:21.968843] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:00.306 [2024-09-29 16:45:21.969037] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:00.306 [2024-09-29 16:45:21.969072] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:00.306 [2024-09-29 16:45:21.969338] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:00.306 BaseBdev2 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.306 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.565 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.565 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:00.565 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.565 16:45:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.565 [ 00:07:00.565 { 00:07:00.565 "name": "BaseBdev2", 00:07:00.565 "aliases": [ 00:07:00.566 "5f1f0607-b95d-4f76-936c-b4aa055cc083" 00:07:00.566 ], 00:07:00.566 "product_name": "Malloc disk", 00:07:00.566 "block_size": 512, 00:07:00.566 "num_blocks": 65536, 00:07:00.566 "uuid": "5f1f0607-b95d-4f76-936c-b4aa055cc083", 00:07:00.566 "assigned_rate_limits": { 00:07:00.566 "rw_ios_per_sec": 0, 00:07:00.566 "rw_mbytes_per_sec": 0, 00:07:00.566 "r_mbytes_per_sec": 0, 00:07:00.566 "w_mbytes_per_sec": 0 00:07:00.566 }, 00:07:00.566 "claimed": true, 00:07:00.566 "claim_type": "exclusive_write", 00:07:00.566 "zoned": false, 00:07:00.566 "supported_io_types": { 00:07:00.566 "read": true, 00:07:00.566 "write": true, 00:07:00.566 "unmap": true, 00:07:00.566 "flush": true, 00:07:00.566 "reset": true, 00:07:00.566 "nvme_admin": false, 00:07:00.566 "nvme_io": false, 00:07:00.566 "nvme_io_md": false, 00:07:00.566 "write_zeroes": true, 00:07:00.566 "zcopy": true, 00:07:00.566 "get_zone_info": false, 00:07:00.566 "zone_management": false, 00:07:00.566 "zone_append": false, 00:07:00.566 "compare": false, 00:07:00.566 "compare_and_write": false, 00:07:00.566 "abort": true, 00:07:00.566 "seek_hole": false, 00:07:00.566 "seek_data": false, 00:07:00.566 "copy": true, 00:07:00.566 "nvme_iov_md": false 00:07:00.566 }, 00:07:00.566 "memory_domains": [ 00:07:00.566 { 00:07:00.566 "dma_device_id": "system", 00:07:00.566 "dma_device_type": 1 00:07:00.566 }, 00:07:00.566 { 00:07:00.566 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:00.566 "dma_device_type": 2 00:07:00.566 } 00:07:00.566 ], 00:07:00.566 "driver_specific": {} 00:07:00.566 } 00:07:00.566 ] 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:00.566 "name": "Existed_Raid", 00:07:00.566 "uuid": "d67682a2-d6f7-4085-927a-d5d6db933423", 00:07:00.566 "strip_size_kb": 64, 00:07:00.566 "state": "online", 00:07:00.566 "raid_level": "raid0", 00:07:00.566 "superblock": false, 00:07:00.566 "num_base_bdevs": 2, 00:07:00.566 "num_base_bdevs_discovered": 2, 00:07:00.566 "num_base_bdevs_operational": 2, 00:07:00.566 "base_bdevs_list": [ 00:07:00.566 { 00:07:00.566 "name": "BaseBdev1", 00:07:00.566 "uuid": "718a3e8e-6a19-44e2-bbaa-9a79ef2f6ac7", 00:07:00.566 "is_configured": true, 00:07:00.566 "data_offset": 0, 00:07:00.566 "data_size": 65536 00:07:00.566 }, 00:07:00.566 { 00:07:00.566 "name": "BaseBdev2", 00:07:00.566 "uuid": "5f1f0607-b95d-4f76-936c-b4aa055cc083", 00:07:00.566 "is_configured": true, 00:07:00.566 "data_offset": 0, 00:07:00.566 "data_size": 65536 00:07:00.566 } 00:07:00.566 ] 00:07:00.566 }' 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:00.566 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:00.825 [2024-09-29 16:45:22.423881] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:00.825 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:00.825 "name": "Existed_Raid", 00:07:00.825 "aliases": [ 00:07:00.825 "d67682a2-d6f7-4085-927a-d5d6db933423" 00:07:00.825 ], 00:07:00.825 "product_name": "Raid Volume", 00:07:00.825 "block_size": 512, 00:07:00.825 "num_blocks": 131072, 00:07:00.825 "uuid": "d67682a2-d6f7-4085-927a-d5d6db933423", 00:07:00.825 "assigned_rate_limits": { 00:07:00.825 "rw_ios_per_sec": 0, 00:07:00.825 "rw_mbytes_per_sec": 0, 00:07:00.825 "r_mbytes_per_sec": 0, 00:07:00.825 "w_mbytes_per_sec": 0 00:07:00.825 }, 00:07:00.825 "claimed": false, 00:07:00.825 "zoned": false, 00:07:00.825 "supported_io_types": { 00:07:00.825 "read": true, 00:07:00.825 "write": true, 00:07:00.825 "unmap": true, 00:07:00.825 "flush": true, 00:07:00.825 "reset": true, 00:07:00.825 "nvme_admin": false, 00:07:00.825 "nvme_io": false, 00:07:00.825 "nvme_io_md": false, 00:07:00.825 "write_zeroes": true, 00:07:00.825 "zcopy": false, 00:07:00.825 "get_zone_info": false, 00:07:00.825 "zone_management": false, 00:07:00.825 "zone_append": false, 00:07:00.825 "compare": false, 00:07:00.825 "compare_and_write": false, 00:07:00.825 "abort": false, 00:07:00.825 "seek_hole": false, 00:07:00.825 "seek_data": false, 00:07:00.825 "copy": false, 00:07:00.826 "nvme_iov_md": false 00:07:00.826 }, 00:07:00.826 "memory_domains": [ 00:07:00.826 { 00:07:00.826 "dma_device_id": "system", 00:07:00.826 "dma_device_type": 1 00:07:00.826 }, 00:07:00.826 { 00:07:00.826 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:00.826 "dma_device_type": 2 00:07:00.826 }, 00:07:00.826 { 00:07:00.826 "dma_device_id": "system", 00:07:00.826 "dma_device_type": 1 00:07:00.826 }, 00:07:00.826 { 00:07:00.826 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:00.826 "dma_device_type": 2 00:07:00.826 } 00:07:00.826 ], 00:07:00.826 "driver_specific": { 00:07:00.826 "raid": { 00:07:00.826 "uuid": "d67682a2-d6f7-4085-927a-d5d6db933423", 00:07:00.826 "strip_size_kb": 64, 00:07:00.826 "state": "online", 00:07:00.826 "raid_level": "raid0", 00:07:00.826 "superblock": false, 00:07:00.826 "num_base_bdevs": 2, 00:07:00.826 "num_base_bdevs_discovered": 2, 00:07:00.826 "num_base_bdevs_operational": 2, 00:07:00.826 "base_bdevs_list": [ 00:07:00.826 { 00:07:00.826 "name": "BaseBdev1", 00:07:00.826 "uuid": "718a3e8e-6a19-44e2-bbaa-9a79ef2f6ac7", 00:07:00.826 "is_configured": true, 00:07:00.826 "data_offset": 0, 00:07:00.826 "data_size": 65536 00:07:00.826 }, 00:07:00.826 { 00:07:00.826 "name": "BaseBdev2", 00:07:00.826 "uuid": "5f1f0607-b95d-4f76-936c-b4aa055cc083", 00:07:00.826 "is_configured": true, 00:07:00.826 "data_offset": 0, 00:07:00.826 "data_size": 65536 00:07:00.826 } 00:07:00.826 ] 00:07:00.826 } 00:07:00.826 } 00:07:00.826 }' 00:07:00.826 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:01.084 BaseBdev2' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.084 [2024-09-29 16:45:22.603405] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:01.084 [2024-09-29 16:45:22.603436] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:01.084 [2024-09-29 16:45:22.603489] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:01.084 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.085 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.085 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.085 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.085 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:01.085 "name": "Existed_Raid", 00:07:01.085 "uuid": "d67682a2-d6f7-4085-927a-d5d6db933423", 00:07:01.085 "strip_size_kb": 64, 00:07:01.085 "state": "offline", 00:07:01.085 "raid_level": "raid0", 00:07:01.085 "superblock": false, 00:07:01.085 "num_base_bdevs": 2, 00:07:01.085 "num_base_bdevs_discovered": 1, 00:07:01.085 "num_base_bdevs_operational": 1, 00:07:01.085 "base_bdevs_list": [ 00:07:01.085 { 00:07:01.085 "name": null, 00:07:01.085 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:01.085 "is_configured": false, 00:07:01.085 "data_offset": 0, 00:07:01.085 "data_size": 65536 00:07:01.085 }, 00:07:01.085 { 00:07:01.085 "name": "BaseBdev2", 00:07:01.085 "uuid": "5f1f0607-b95d-4f76-936c-b4aa055cc083", 00:07:01.085 "is_configured": true, 00:07:01.085 "data_offset": 0, 00:07:01.085 "data_size": 65536 00:07:01.085 } 00:07:01.085 ] 00:07:01.085 }' 00:07:01.085 16:45:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:01.085 16:45:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.655 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.655 [2024-09-29 16:45:23.081940] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:01.655 [2024-09-29 16:45:23.082007] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 71817 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 71817 ']' 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 71817 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71817 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:01.656 killing process with pid 71817 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71817' 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 71817 00:07:01.656 [2024-09-29 16:45:23.186315] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:01.656 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 71817 00:07:01.656 [2024-09-29 16:45:23.187309] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:01.917 00:07:01.917 real 0m3.759s 00:07:01.917 user 0m5.920s 00:07:01.917 sys 0m0.733s 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:01.917 ************************************ 00:07:01.917 END TEST raid_state_function_test 00:07:01.917 ************************************ 00:07:01.917 16:45:23 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:07:01.917 16:45:23 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:01.917 16:45:23 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:01.917 16:45:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:01.917 ************************************ 00:07:01.917 START TEST raid_state_function_test_sb 00:07:01.917 ************************************ 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 true 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72054 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72054' 00:07:01.917 Process raid pid: 72054 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72054 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 72054 ']' 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:01.917 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:01.917 16:45:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:01.917 [2024-09-29 16:45:23.584806] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:01.917 [2024-09-29 16:45:23.584915] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:02.177 [2024-09-29 16:45:23.711099] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:02.177 [2024-09-29 16:45:23.758120] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:02.177 [2024-09-29 16:45:23.800543] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:02.177 [2024-09-29 16:45:23.800584] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:02.770 [2024-09-29 16:45:24.406003] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:02.770 [2024-09-29 16:45:24.406051] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:02.770 [2024-09-29 16:45:24.406063] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:02.770 [2024-09-29 16:45:24.406072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:02.770 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:02.771 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.044 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:03.044 "name": "Existed_Raid", 00:07:03.044 "uuid": "5e434e4d-69d8-4dc9-aa93-08689a5ba2e8", 00:07:03.044 "strip_size_kb": 64, 00:07:03.044 "state": "configuring", 00:07:03.044 "raid_level": "raid0", 00:07:03.044 "superblock": true, 00:07:03.044 "num_base_bdevs": 2, 00:07:03.044 "num_base_bdevs_discovered": 0, 00:07:03.044 "num_base_bdevs_operational": 2, 00:07:03.044 "base_bdevs_list": [ 00:07:03.044 { 00:07:03.044 "name": "BaseBdev1", 00:07:03.044 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:03.044 "is_configured": false, 00:07:03.044 "data_offset": 0, 00:07:03.044 "data_size": 0 00:07:03.044 }, 00:07:03.044 { 00:07:03.044 "name": "BaseBdev2", 00:07:03.044 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:03.044 "is_configured": false, 00:07:03.044 "data_offset": 0, 00:07:03.044 "data_size": 0 00:07:03.044 } 00:07:03.044 ] 00:07:03.044 }' 00:07:03.044 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:03.044 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 [2024-09-29 16:45:24.801248] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:03.304 [2024-09-29 16:45:24.801294] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 [2024-09-29 16:45:24.813225] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:03.304 [2024-09-29 16:45:24.813262] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:03.304 [2024-09-29 16:45:24.813280] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:03.304 [2024-09-29 16:45:24.813290] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 [2024-09-29 16:45:24.834016] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:03.304 BaseBdev1 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.304 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.304 [ 00:07:03.304 { 00:07:03.304 "name": "BaseBdev1", 00:07:03.304 "aliases": [ 00:07:03.304 "722cea10-6c27-463b-bd21-61b8d39c8f2b" 00:07:03.304 ], 00:07:03.304 "product_name": "Malloc disk", 00:07:03.304 "block_size": 512, 00:07:03.304 "num_blocks": 65536, 00:07:03.304 "uuid": "722cea10-6c27-463b-bd21-61b8d39c8f2b", 00:07:03.304 "assigned_rate_limits": { 00:07:03.304 "rw_ios_per_sec": 0, 00:07:03.304 "rw_mbytes_per_sec": 0, 00:07:03.304 "r_mbytes_per_sec": 0, 00:07:03.304 "w_mbytes_per_sec": 0 00:07:03.304 }, 00:07:03.304 "claimed": true, 00:07:03.304 "claim_type": "exclusive_write", 00:07:03.304 "zoned": false, 00:07:03.304 "supported_io_types": { 00:07:03.304 "read": true, 00:07:03.304 "write": true, 00:07:03.304 "unmap": true, 00:07:03.304 "flush": true, 00:07:03.304 "reset": true, 00:07:03.304 "nvme_admin": false, 00:07:03.304 "nvme_io": false, 00:07:03.304 "nvme_io_md": false, 00:07:03.304 "write_zeroes": true, 00:07:03.304 "zcopy": true, 00:07:03.305 "get_zone_info": false, 00:07:03.305 "zone_management": false, 00:07:03.305 "zone_append": false, 00:07:03.305 "compare": false, 00:07:03.305 "compare_and_write": false, 00:07:03.305 "abort": true, 00:07:03.305 "seek_hole": false, 00:07:03.305 "seek_data": false, 00:07:03.305 "copy": true, 00:07:03.305 "nvme_iov_md": false 00:07:03.305 }, 00:07:03.305 "memory_domains": [ 00:07:03.305 { 00:07:03.305 "dma_device_id": "system", 00:07:03.305 "dma_device_type": 1 00:07:03.305 }, 00:07:03.305 { 00:07:03.305 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:03.305 "dma_device_type": 2 00:07:03.305 } 00:07:03.305 ], 00:07:03.305 "driver_specific": {} 00:07:03.305 } 00:07:03.305 ] 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:03.305 "name": "Existed_Raid", 00:07:03.305 "uuid": "85f17ec9-5673-4c4e-a5b9-238fcfa0619c", 00:07:03.305 "strip_size_kb": 64, 00:07:03.305 "state": "configuring", 00:07:03.305 "raid_level": "raid0", 00:07:03.305 "superblock": true, 00:07:03.305 "num_base_bdevs": 2, 00:07:03.305 "num_base_bdevs_discovered": 1, 00:07:03.305 "num_base_bdevs_operational": 2, 00:07:03.305 "base_bdevs_list": [ 00:07:03.305 { 00:07:03.305 "name": "BaseBdev1", 00:07:03.305 "uuid": "722cea10-6c27-463b-bd21-61b8d39c8f2b", 00:07:03.305 "is_configured": true, 00:07:03.305 "data_offset": 2048, 00:07:03.305 "data_size": 63488 00:07:03.305 }, 00:07:03.305 { 00:07:03.305 "name": "BaseBdev2", 00:07:03.305 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:03.305 "is_configured": false, 00:07:03.305 "data_offset": 0, 00:07:03.305 "data_size": 0 00:07:03.305 } 00:07:03.305 ] 00:07:03.305 }' 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:03.305 16:45:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.872 [2024-09-29 16:45:25.301234] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:03.872 [2024-09-29 16:45:25.301280] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.872 [2024-09-29 16:45:25.313260] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:03.872 [2024-09-29 16:45:25.315091] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:03.872 [2024-09-29 16:45:25.315125] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.872 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:03.872 "name": "Existed_Raid", 00:07:03.872 "uuid": "0a88723a-87a6-4a34-9af2-91fa1e64802f", 00:07:03.872 "strip_size_kb": 64, 00:07:03.872 "state": "configuring", 00:07:03.872 "raid_level": "raid0", 00:07:03.872 "superblock": true, 00:07:03.872 "num_base_bdevs": 2, 00:07:03.872 "num_base_bdevs_discovered": 1, 00:07:03.872 "num_base_bdevs_operational": 2, 00:07:03.872 "base_bdevs_list": [ 00:07:03.872 { 00:07:03.872 "name": "BaseBdev1", 00:07:03.872 "uuid": "722cea10-6c27-463b-bd21-61b8d39c8f2b", 00:07:03.872 "is_configured": true, 00:07:03.872 "data_offset": 2048, 00:07:03.872 "data_size": 63488 00:07:03.872 }, 00:07:03.872 { 00:07:03.872 "name": "BaseBdev2", 00:07:03.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:03.873 "is_configured": false, 00:07:03.873 "data_offset": 0, 00:07:03.873 "data_size": 0 00:07:03.873 } 00:07:03.873 ] 00:07:03.873 }' 00:07:03.873 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:03.873 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.133 [2024-09-29 16:45:25.709862] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:04.133 [2024-09-29 16:45:25.710397] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:04.133 [2024-09-29 16:45:25.710484] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:04.133 BaseBdev2 00:07:04.133 [2024-09-29 16:45:25.711415] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.133 [2024-09-29 16:45:25.711927] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:04.133 [2024-09-29 16:45:25.712021] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:04.133 [2024-09-29 16:45:25.712447] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.133 [ 00:07:04.133 { 00:07:04.133 "name": "BaseBdev2", 00:07:04.133 "aliases": [ 00:07:04.133 "3736c0a8-332a-4c7c-b871-02dd5aa718a6" 00:07:04.133 ], 00:07:04.133 "product_name": "Malloc disk", 00:07:04.133 "block_size": 512, 00:07:04.133 "num_blocks": 65536, 00:07:04.133 "uuid": "3736c0a8-332a-4c7c-b871-02dd5aa718a6", 00:07:04.133 "assigned_rate_limits": { 00:07:04.133 "rw_ios_per_sec": 0, 00:07:04.133 "rw_mbytes_per_sec": 0, 00:07:04.133 "r_mbytes_per_sec": 0, 00:07:04.133 "w_mbytes_per_sec": 0 00:07:04.133 }, 00:07:04.133 "claimed": true, 00:07:04.133 "claim_type": "exclusive_write", 00:07:04.133 "zoned": false, 00:07:04.133 "supported_io_types": { 00:07:04.133 "read": true, 00:07:04.133 "write": true, 00:07:04.133 "unmap": true, 00:07:04.133 "flush": true, 00:07:04.133 "reset": true, 00:07:04.133 "nvme_admin": false, 00:07:04.133 "nvme_io": false, 00:07:04.133 "nvme_io_md": false, 00:07:04.133 "write_zeroes": true, 00:07:04.133 "zcopy": true, 00:07:04.133 "get_zone_info": false, 00:07:04.133 "zone_management": false, 00:07:04.133 "zone_append": false, 00:07:04.133 "compare": false, 00:07:04.133 "compare_and_write": false, 00:07:04.133 "abort": true, 00:07:04.133 "seek_hole": false, 00:07:04.133 "seek_data": false, 00:07:04.133 "copy": true, 00:07:04.133 "nvme_iov_md": false 00:07:04.133 }, 00:07:04.133 "memory_domains": [ 00:07:04.133 { 00:07:04.133 "dma_device_id": "system", 00:07:04.133 "dma_device_type": 1 00:07:04.133 }, 00:07:04.133 { 00:07:04.133 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:04.133 "dma_device_type": 2 00:07:04.133 } 00:07:04.133 ], 00:07:04.133 "driver_specific": {} 00:07:04.133 } 00:07:04.133 ] 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:04.133 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:04.134 "name": "Existed_Raid", 00:07:04.134 "uuid": "0a88723a-87a6-4a34-9af2-91fa1e64802f", 00:07:04.134 "strip_size_kb": 64, 00:07:04.134 "state": "online", 00:07:04.134 "raid_level": "raid0", 00:07:04.134 "superblock": true, 00:07:04.134 "num_base_bdevs": 2, 00:07:04.134 "num_base_bdevs_discovered": 2, 00:07:04.134 "num_base_bdevs_operational": 2, 00:07:04.134 "base_bdevs_list": [ 00:07:04.134 { 00:07:04.134 "name": "BaseBdev1", 00:07:04.134 "uuid": "722cea10-6c27-463b-bd21-61b8d39c8f2b", 00:07:04.134 "is_configured": true, 00:07:04.134 "data_offset": 2048, 00:07:04.134 "data_size": 63488 00:07:04.134 }, 00:07:04.134 { 00:07:04.134 "name": "BaseBdev2", 00:07:04.134 "uuid": "3736c0a8-332a-4c7c-b871-02dd5aa718a6", 00:07:04.134 "is_configured": true, 00:07:04.134 "data_offset": 2048, 00:07:04.134 "data_size": 63488 00:07:04.134 } 00:07:04.134 ] 00:07:04.134 }' 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:04.134 16:45:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:04.705 [2024-09-29 16:45:26.225170] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:04.705 "name": "Existed_Raid", 00:07:04.705 "aliases": [ 00:07:04.705 "0a88723a-87a6-4a34-9af2-91fa1e64802f" 00:07:04.705 ], 00:07:04.705 "product_name": "Raid Volume", 00:07:04.705 "block_size": 512, 00:07:04.705 "num_blocks": 126976, 00:07:04.705 "uuid": "0a88723a-87a6-4a34-9af2-91fa1e64802f", 00:07:04.705 "assigned_rate_limits": { 00:07:04.705 "rw_ios_per_sec": 0, 00:07:04.705 "rw_mbytes_per_sec": 0, 00:07:04.705 "r_mbytes_per_sec": 0, 00:07:04.705 "w_mbytes_per_sec": 0 00:07:04.705 }, 00:07:04.705 "claimed": false, 00:07:04.705 "zoned": false, 00:07:04.705 "supported_io_types": { 00:07:04.705 "read": true, 00:07:04.705 "write": true, 00:07:04.705 "unmap": true, 00:07:04.705 "flush": true, 00:07:04.705 "reset": true, 00:07:04.705 "nvme_admin": false, 00:07:04.705 "nvme_io": false, 00:07:04.705 "nvme_io_md": false, 00:07:04.705 "write_zeroes": true, 00:07:04.705 "zcopy": false, 00:07:04.705 "get_zone_info": false, 00:07:04.705 "zone_management": false, 00:07:04.705 "zone_append": false, 00:07:04.705 "compare": false, 00:07:04.705 "compare_and_write": false, 00:07:04.705 "abort": false, 00:07:04.705 "seek_hole": false, 00:07:04.705 "seek_data": false, 00:07:04.705 "copy": false, 00:07:04.705 "nvme_iov_md": false 00:07:04.705 }, 00:07:04.705 "memory_domains": [ 00:07:04.705 { 00:07:04.705 "dma_device_id": "system", 00:07:04.705 "dma_device_type": 1 00:07:04.705 }, 00:07:04.705 { 00:07:04.705 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:04.705 "dma_device_type": 2 00:07:04.705 }, 00:07:04.705 { 00:07:04.705 "dma_device_id": "system", 00:07:04.705 "dma_device_type": 1 00:07:04.705 }, 00:07:04.705 { 00:07:04.705 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:04.705 "dma_device_type": 2 00:07:04.705 } 00:07:04.705 ], 00:07:04.705 "driver_specific": { 00:07:04.705 "raid": { 00:07:04.705 "uuid": "0a88723a-87a6-4a34-9af2-91fa1e64802f", 00:07:04.705 "strip_size_kb": 64, 00:07:04.705 "state": "online", 00:07:04.705 "raid_level": "raid0", 00:07:04.705 "superblock": true, 00:07:04.705 "num_base_bdevs": 2, 00:07:04.705 "num_base_bdevs_discovered": 2, 00:07:04.705 "num_base_bdevs_operational": 2, 00:07:04.705 "base_bdevs_list": [ 00:07:04.705 { 00:07:04.705 "name": "BaseBdev1", 00:07:04.705 "uuid": "722cea10-6c27-463b-bd21-61b8d39c8f2b", 00:07:04.705 "is_configured": true, 00:07:04.705 "data_offset": 2048, 00:07:04.705 "data_size": 63488 00:07:04.705 }, 00:07:04.705 { 00:07:04.705 "name": "BaseBdev2", 00:07:04.705 "uuid": "3736c0a8-332a-4c7c-b871-02dd5aa718a6", 00:07:04.705 "is_configured": true, 00:07:04.705 "data_offset": 2048, 00:07:04.705 "data_size": 63488 00:07:04.705 } 00:07:04.705 ] 00:07:04.705 } 00:07:04.705 } 00:07:04.705 }' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:04.705 BaseBdev2' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:04.705 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.965 [2024-09-29 16:45:26.412619] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:04.965 [2024-09-29 16:45:26.412649] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:04.965 [2024-09-29 16:45:26.412699] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:04.965 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:04.965 "name": "Existed_Raid", 00:07:04.965 "uuid": "0a88723a-87a6-4a34-9af2-91fa1e64802f", 00:07:04.965 "strip_size_kb": 64, 00:07:04.965 "state": "offline", 00:07:04.965 "raid_level": "raid0", 00:07:04.966 "superblock": true, 00:07:04.966 "num_base_bdevs": 2, 00:07:04.966 "num_base_bdevs_discovered": 1, 00:07:04.966 "num_base_bdevs_operational": 1, 00:07:04.966 "base_bdevs_list": [ 00:07:04.966 { 00:07:04.966 "name": null, 00:07:04.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:04.966 "is_configured": false, 00:07:04.966 "data_offset": 0, 00:07:04.966 "data_size": 63488 00:07:04.966 }, 00:07:04.966 { 00:07:04.966 "name": "BaseBdev2", 00:07:04.966 "uuid": "3736c0a8-332a-4c7c-b871-02dd5aa718a6", 00:07:04.966 "is_configured": true, 00:07:04.966 "data_offset": 2048, 00:07:04.966 "data_size": 63488 00:07:04.966 } 00:07:04.966 ] 00:07:04.966 }' 00:07:04.966 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:04.966 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.225 [2024-09-29 16:45:26.879173] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:05.225 [2024-09-29 16:45:26.879236] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:05.225 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:05.226 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:05.226 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.226 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72054 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 72054 ']' 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 72054 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72054 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:05.486 killing process with pid 72054 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72054' 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 72054 00:07:05.486 [2024-09-29 16:45:26.983228] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:05.486 16:45:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 72054 00:07:05.486 [2024-09-29 16:45:26.984207] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:05.746 16:45:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:05.746 00:07:05.746 real 0m3.727s 00:07:05.746 user 0m5.870s 00:07:05.746 sys 0m0.688s 00:07:05.746 16:45:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:05.746 16:45:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:05.746 ************************************ 00:07:05.746 END TEST raid_state_function_test_sb 00:07:05.746 ************************************ 00:07:05.746 16:45:27 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:07:05.746 16:45:27 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:05.746 16:45:27 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:05.746 16:45:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:05.746 ************************************ 00:07:05.746 START TEST raid_superblock_test 00:07:05.746 ************************************ 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 2 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72295 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72295 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 72295 ']' 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:05.746 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:05.746 16:45:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:05.746 [2024-09-29 16:45:27.378924] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:05.747 [2024-09-29 16:45:27.379060] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72295 ] 00:07:06.005 [2024-09-29 16:45:27.523101] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:06.005 [2024-09-29 16:45:27.568314] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:06.006 [2024-09-29 16:45:27.611344] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:06.006 [2024-09-29 16:45:27.611388] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:06.574 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:06.574 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.575 malloc1 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.575 [2024-09-29 16:45:28.217847] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:06.575 [2024-09-29 16:45:28.217899] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.575 [2024-09-29 16:45:28.217929] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:06.575 [2024-09-29 16:45:28.217945] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.575 [2024-09-29 16:45:28.220053] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.575 [2024-09-29 16:45:28.220091] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:06.575 pt1 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.575 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.835 malloc2 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.835 [2024-09-29 16:45:28.260596] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:06.835 [2024-09-29 16:45:28.260764] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:06.835 [2024-09-29 16:45:28.260822] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:06.835 [2024-09-29 16:45:28.260859] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:06.835 [2024-09-29 16:45:28.265279] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:06.835 [2024-09-29 16:45:28.265336] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:06.835 pt2 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.835 [2024-09-29 16:45:28.273590] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:06.835 [2024-09-29 16:45:28.276110] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:06.835 [2024-09-29 16:45:28.276289] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:06.835 [2024-09-29 16:45:28.276316] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:06.835 [2024-09-29 16:45:28.276684] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:06.835 [2024-09-29 16:45:28.276901] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:06.835 [2024-09-29 16:45:28.276937] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:06.835 [2024-09-29 16:45:28.277103] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:06.835 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:06.836 "name": "raid_bdev1", 00:07:06.836 "uuid": "22960249-fd86-4aff-bef6-75ef1193a5c4", 00:07:06.836 "strip_size_kb": 64, 00:07:06.836 "state": "online", 00:07:06.836 "raid_level": "raid0", 00:07:06.836 "superblock": true, 00:07:06.836 "num_base_bdevs": 2, 00:07:06.836 "num_base_bdevs_discovered": 2, 00:07:06.836 "num_base_bdevs_operational": 2, 00:07:06.836 "base_bdevs_list": [ 00:07:06.836 { 00:07:06.836 "name": "pt1", 00:07:06.836 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:06.836 "is_configured": true, 00:07:06.836 "data_offset": 2048, 00:07:06.836 "data_size": 63488 00:07:06.836 }, 00:07:06.836 { 00:07:06.836 "name": "pt2", 00:07:06.836 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:06.836 "is_configured": true, 00:07:06.836 "data_offset": 2048, 00:07:06.836 "data_size": 63488 00:07:06.836 } 00:07:06.836 ] 00:07:06.836 }' 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:06.836 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.096 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.096 [2024-09-29 16:45:28.756952] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:07.356 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.356 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:07.356 "name": "raid_bdev1", 00:07:07.356 "aliases": [ 00:07:07.356 "22960249-fd86-4aff-bef6-75ef1193a5c4" 00:07:07.356 ], 00:07:07.356 "product_name": "Raid Volume", 00:07:07.356 "block_size": 512, 00:07:07.356 "num_blocks": 126976, 00:07:07.356 "uuid": "22960249-fd86-4aff-bef6-75ef1193a5c4", 00:07:07.356 "assigned_rate_limits": { 00:07:07.356 "rw_ios_per_sec": 0, 00:07:07.356 "rw_mbytes_per_sec": 0, 00:07:07.356 "r_mbytes_per_sec": 0, 00:07:07.356 "w_mbytes_per_sec": 0 00:07:07.356 }, 00:07:07.356 "claimed": false, 00:07:07.356 "zoned": false, 00:07:07.356 "supported_io_types": { 00:07:07.356 "read": true, 00:07:07.356 "write": true, 00:07:07.356 "unmap": true, 00:07:07.356 "flush": true, 00:07:07.356 "reset": true, 00:07:07.356 "nvme_admin": false, 00:07:07.356 "nvme_io": false, 00:07:07.356 "nvme_io_md": false, 00:07:07.356 "write_zeroes": true, 00:07:07.356 "zcopy": false, 00:07:07.356 "get_zone_info": false, 00:07:07.356 "zone_management": false, 00:07:07.356 "zone_append": false, 00:07:07.356 "compare": false, 00:07:07.356 "compare_and_write": false, 00:07:07.356 "abort": false, 00:07:07.357 "seek_hole": false, 00:07:07.357 "seek_data": false, 00:07:07.357 "copy": false, 00:07:07.357 "nvme_iov_md": false 00:07:07.357 }, 00:07:07.357 "memory_domains": [ 00:07:07.357 { 00:07:07.357 "dma_device_id": "system", 00:07:07.357 "dma_device_type": 1 00:07:07.357 }, 00:07:07.357 { 00:07:07.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:07.357 "dma_device_type": 2 00:07:07.357 }, 00:07:07.357 { 00:07:07.357 "dma_device_id": "system", 00:07:07.357 "dma_device_type": 1 00:07:07.357 }, 00:07:07.357 { 00:07:07.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:07.357 "dma_device_type": 2 00:07:07.357 } 00:07:07.357 ], 00:07:07.357 "driver_specific": { 00:07:07.357 "raid": { 00:07:07.357 "uuid": "22960249-fd86-4aff-bef6-75ef1193a5c4", 00:07:07.357 "strip_size_kb": 64, 00:07:07.357 "state": "online", 00:07:07.357 "raid_level": "raid0", 00:07:07.357 "superblock": true, 00:07:07.357 "num_base_bdevs": 2, 00:07:07.357 "num_base_bdevs_discovered": 2, 00:07:07.357 "num_base_bdevs_operational": 2, 00:07:07.357 "base_bdevs_list": [ 00:07:07.357 { 00:07:07.357 "name": "pt1", 00:07:07.357 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:07.357 "is_configured": true, 00:07:07.357 "data_offset": 2048, 00:07:07.357 "data_size": 63488 00:07:07.357 }, 00:07:07.357 { 00:07:07.357 "name": "pt2", 00:07:07.357 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:07.357 "is_configured": true, 00:07:07.357 "data_offset": 2048, 00:07:07.357 "data_size": 63488 00:07:07.357 } 00:07:07.357 ] 00:07:07.357 } 00:07:07.357 } 00:07:07.357 }' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:07.357 pt2' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:07.357 [2024-09-29 16:45:28.968501] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=22960249-fd86-4aff-bef6-75ef1193a5c4 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 22960249-fd86-4aff-bef6-75ef1193a5c4 ']' 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.357 16:45:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.357 [2024-09-29 16:45:29.004230] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:07.357 [2024-09-29 16:45:29.004261] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:07.357 [2024-09-29 16:45:29.004338] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:07.357 [2024-09-29 16:45:29.004383] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:07.357 [2024-09-29 16:45:29.004393] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:07.357 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.357 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:07.357 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.357 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.357 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:07.357 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.616 [2024-09-29 16:45:29.136021] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:07.616 [2024-09-29 16:45:29.137929] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:07.616 [2024-09-29 16:45:29.138011] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:07.616 [2024-09-29 16:45:29.138048] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:07.616 [2024-09-29 16:45:29.138064] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:07.616 [2024-09-29 16:45:29.138073] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:07.616 request: 00:07:07.616 { 00:07:07.616 "name": "raid_bdev1", 00:07:07.616 "raid_level": "raid0", 00:07:07.616 "base_bdevs": [ 00:07:07.616 "malloc1", 00:07:07.616 "malloc2" 00:07:07.616 ], 00:07:07.616 "strip_size_kb": 64, 00:07:07.616 "superblock": false, 00:07:07.616 "method": "bdev_raid_create", 00:07:07.616 "req_id": 1 00:07:07.616 } 00:07:07.616 Got JSON-RPC error response 00:07:07.616 response: 00:07:07.616 { 00:07:07.616 "code": -17, 00:07:07.616 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:07.616 } 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:07.616 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.617 [2024-09-29 16:45:29.199877] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:07.617 [2024-09-29 16:45:29.199928] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:07.617 [2024-09-29 16:45:29.199949] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:07.617 [2024-09-29 16:45:29.199957] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:07.617 [2024-09-29 16:45:29.202112] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:07.617 [2024-09-29 16:45:29.202141] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:07.617 [2024-09-29 16:45:29.202221] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:07.617 [2024-09-29 16:45:29.202250] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:07.617 pt1 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:07.617 "name": "raid_bdev1", 00:07:07.617 "uuid": "22960249-fd86-4aff-bef6-75ef1193a5c4", 00:07:07.617 "strip_size_kb": 64, 00:07:07.617 "state": "configuring", 00:07:07.617 "raid_level": "raid0", 00:07:07.617 "superblock": true, 00:07:07.617 "num_base_bdevs": 2, 00:07:07.617 "num_base_bdevs_discovered": 1, 00:07:07.617 "num_base_bdevs_operational": 2, 00:07:07.617 "base_bdevs_list": [ 00:07:07.617 { 00:07:07.617 "name": "pt1", 00:07:07.617 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:07.617 "is_configured": true, 00:07:07.617 "data_offset": 2048, 00:07:07.617 "data_size": 63488 00:07:07.617 }, 00:07:07.617 { 00:07:07.617 "name": null, 00:07:07.617 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:07.617 "is_configured": false, 00:07:07.617 "data_offset": 2048, 00:07:07.617 "data_size": 63488 00:07:07.617 } 00:07:07.617 ] 00:07:07.617 }' 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:07.617 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.184 [2024-09-29 16:45:29.587273] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:08.184 [2024-09-29 16:45:29.587320] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:08.184 [2024-09-29 16:45:29.587339] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:08.184 [2024-09-29 16:45:29.587347] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:08.184 [2024-09-29 16:45:29.587714] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:08.184 [2024-09-29 16:45:29.587744] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:08.184 [2024-09-29 16:45:29.587809] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:08.184 [2024-09-29 16:45:29.587828] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:08.184 [2024-09-29 16:45:29.587907] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:08.184 [2024-09-29 16:45:29.587916] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:08.184 [2024-09-29 16:45:29.588147] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:08.184 [2024-09-29 16:45:29.588244] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:08.184 [2024-09-29 16:45:29.588257] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:08.184 [2024-09-29 16:45:29.588350] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:08.184 pt2 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:08.184 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:08.185 "name": "raid_bdev1", 00:07:08.185 "uuid": "22960249-fd86-4aff-bef6-75ef1193a5c4", 00:07:08.185 "strip_size_kb": 64, 00:07:08.185 "state": "online", 00:07:08.185 "raid_level": "raid0", 00:07:08.185 "superblock": true, 00:07:08.185 "num_base_bdevs": 2, 00:07:08.185 "num_base_bdevs_discovered": 2, 00:07:08.185 "num_base_bdevs_operational": 2, 00:07:08.185 "base_bdevs_list": [ 00:07:08.185 { 00:07:08.185 "name": "pt1", 00:07:08.185 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:08.185 "is_configured": true, 00:07:08.185 "data_offset": 2048, 00:07:08.185 "data_size": 63488 00:07:08.185 }, 00:07:08.185 { 00:07:08.185 "name": "pt2", 00:07:08.185 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:08.185 "is_configured": true, 00:07:08.185 "data_offset": 2048, 00:07:08.185 "data_size": 63488 00:07:08.185 } 00:07:08.185 ] 00:07:08.185 }' 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:08.185 16:45:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:08.445 [2024-09-29 16:45:30.038798] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:08.445 "name": "raid_bdev1", 00:07:08.445 "aliases": [ 00:07:08.445 "22960249-fd86-4aff-bef6-75ef1193a5c4" 00:07:08.445 ], 00:07:08.445 "product_name": "Raid Volume", 00:07:08.445 "block_size": 512, 00:07:08.445 "num_blocks": 126976, 00:07:08.445 "uuid": "22960249-fd86-4aff-bef6-75ef1193a5c4", 00:07:08.445 "assigned_rate_limits": { 00:07:08.445 "rw_ios_per_sec": 0, 00:07:08.445 "rw_mbytes_per_sec": 0, 00:07:08.445 "r_mbytes_per_sec": 0, 00:07:08.445 "w_mbytes_per_sec": 0 00:07:08.445 }, 00:07:08.445 "claimed": false, 00:07:08.445 "zoned": false, 00:07:08.445 "supported_io_types": { 00:07:08.445 "read": true, 00:07:08.445 "write": true, 00:07:08.445 "unmap": true, 00:07:08.445 "flush": true, 00:07:08.445 "reset": true, 00:07:08.445 "nvme_admin": false, 00:07:08.445 "nvme_io": false, 00:07:08.445 "nvme_io_md": false, 00:07:08.445 "write_zeroes": true, 00:07:08.445 "zcopy": false, 00:07:08.445 "get_zone_info": false, 00:07:08.445 "zone_management": false, 00:07:08.445 "zone_append": false, 00:07:08.445 "compare": false, 00:07:08.445 "compare_and_write": false, 00:07:08.445 "abort": false, 00:07:08.445 "seek_hole": false, 00:07:08.445 "seek_data": false, 00:07:08.445 "copy": false, 00:07:08.445 "nvme_iov_md": false 00:07:08.445 }, 00:07:08.445 "memory_domains": [ 00:07:08.445 { 00:07:08.445 "dma_device_id": "system", 00:07:08.445 "dma_device_type": 1 00:07:08.445 }, 00:07:08.445 { 00:07:08.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:08.445 "dma_device_type": 2 00:07:08.445 }, 00:07:08.445 { 00:07:08.445 "dma_device_id": "system", 00:07:08.445 "dma_device_type": 1 00:07:08.445 }, 00:07:08.445 { 00:07:08.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:08.445 "dma_device_type": 2 00:07:08.445 } 00:07:08.445 ], 00:07:08.445 "driver_specific": { 00:07:08.445 "raid": { 00:07:08.445 "uuid": "22960249-fd86-4aff-bef6-75ef1193a5c4", 00:07:08.445 "strip_size_kb": 64, 00:07:08.445 "state": "online", 00:07:08.445 "raid_level": "raid0", 00:07:08.445 "superblock": true, 00:07:08.445 "num_base_bdevs": 2, 00:07:08.445 "num_base_bdevs_discovered": 2, 00:07:08.445 "num_base_bdevs_operational": 2, 00:07:08.445 "base_bdevs_list": [ 00:07:08.445 { 00:07:08.445 "name": "pt1", 00:07:08.445 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:08.445 "is_configured": true, 00:07:08.445 "data_offset": 2048, 00:07:08.445 "data_size": 63488 00:07:08.445 }, 00:07:08.445 { 00:07:08.445 "name": "pt2", 00:07:08.445 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:08.445 "is_configured": true, 00:07:08.445 "data_offset": 2048, 00:07:08.445 "data_size": 63488 00:07:08.445 } 00:07:08.445 ] 00:07:08.445 } 00:07:08.445 } 00:07:08.445 }' 00:07:08.445 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:08.705 pt2' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.705 [2024-09-29 16:45:30.286366] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 22960249-fd86-4aff-bef6-75ef1193a5c4 '!=' 22960249-fd86-4aff-bef6-75ef1193a5c4 ']' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72295 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 72295 ']' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 72295 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72295 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:08.705 killing process with pid 72295 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72295' 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 72295 00:07:08.705 [2024-09-29 16:45:30.368093] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:08.705 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 72295 00:07:08.705 [2024-09-29 16:45:30.368200] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:08.705 [2024-09-29 16:45:30.368258] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:08.705 [2024-09-29 16:45:30.368272] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:08.965 [2024-09-29 16:45:30.391774] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:08.965 16:45:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:08.965 00:07:08.965 real 0m3.334s 00:07:08.965 user 0m5.169s 00:07:08.965 sys 0m0.671s 00:07:08.965 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:08.965 16:45:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:08.965 ************************************ 00:07:08.965 END TEST raid_superblock_test 00:07:08.965 ************************************ 00:07:09.224 16:45:30 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:07:09.224 16:45:30 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:09.224 16:45:30 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:09.224 16:45:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:09.224 ************************************ 00:07:09.224 START TEST raid_read_error_test 00:07:09.224 ************************************ 00:07:09.224 16:45:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 read 00:07:09.224 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:09.224 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.30mKPUYEYR 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72490 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72490 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 72490 ']' 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:09.225 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:09.225 16:45:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:09.225 [2024-09-29 16:45:30.797011] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:09.225 [2024-09-29 16:45:30.797134] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72490 ] 00:07:09.484 [2024-09-29 16:45:30.941091] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:09.484 [2024-09-29 16:45:30.986194] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:09.484 [2024-09-29 16:45:31.027856] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:09.484 [2024-09-29 16:45:31.027894] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:10.080 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 BaseBdev1_malloc 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 true 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 [2024-09-29 16:45:31.650002] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:10.081 [2024-09-29 16:45:31.650072] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:10.081 [2024-09-29 16:45:31.650100] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:10.081 [2024-09-29 16:45:31.650111] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:10.081 [2024-09-29 16:45:31.652290] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:10.081 [2024-09-29 16:45:31.652324] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:10.081 BaseBdev1 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 BaseBdev2_malloc 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 true 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 [2024-09-29 16:45:31.708910] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:10.081 [2024-09-29 16:45:31.708981] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:10.081 [2024-09-29 16:45:31.709013] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:10.081 [2024-09-29 16:45:31.709027] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:10.081 [2024-09-29 16:45:31.712298] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:10.081 [2024-09-29 16:45:31.712343] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:10.081 BaseBdev2 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 [2024-09-29 16:45:31.721031] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:10.081 [2024-09-29 16:45:31.723104] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:10.081 [2024-09-29 16:45:31.723328] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:10.081 [2024-09-29 16:45:31.723344] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:10.081 [2024-09-29 16:45:31.723629] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:10.081 [2024-09-29 16:45:31.723784] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:10.081 [2024-09-29 16:45:31.723802] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:10.081 [2024-09-29 16:45:31.723980] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.081 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.340 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:10.340 "name": "raid_bdev1", 00:07:10.340 "uuid": "ecebd0f9-7899-49a9-bac7-bb8df6a984ca", 00:07:10.340 "strip_size_kb": 64, 00:07:10.340 "state": "online", 00:07:10.340 "raid_level": "raid0", 00:07:10.340 "superblock": true, 00:07:10.340 "num_base_bdevs": 2, 00:07:10.340 "num_base_bdevs_discovered": 2, 00:07:10.340 "num_base_bdevs_operational": 2, 00:07:10.340 "base_bdevs_list": [ 00:07:10.340 { 00:07:10.340 "name": "BaseBdev1", 00:07:10.340 "uuid": "a128fcfd-fe21-5333-be3a-fc6ad5e9fb41", 00:07:10.340 "is_configured": true, 00:07:10.340 "data_offset": 2048, 00:07:10.340 "data_size": 63488 00:07:10.340 }, 00:07:10.340 { 00:07:10.340 "name": "BaseBdev2", 00:07:10.340 "uuid": "4f014086-3c2b-51a5-95c0-caa4e6766383", 00:07:10.340 "is_configured": true, 00:07:10.341 "data_offset": 2048, 00:07:10.341 "data_size": 63488 00:07:10.341 } 00:07:10.341 ] 00:07:10.341 }' 00:07:10.341 16:45:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:10.341 16:45:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:10.600 16:45:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:10.600 16:45:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:10.600 [2024-09-29 16:45:32.240551] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:11.539 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:11.799 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:11.799 "name": "raid_bdev1", 00:07:11.799 "uuid": "ecebd0f9-7899-49a9-bac7-bb8df6a984ca", 00:07:11.799 "strip_size_kb": 64, 00:07:11.799 "state": "online", 00:07:11.799 "raid_level": "raid0", 00:07:11.799 "superblock": true, 00:07:11.799 "num_base_bdevs": 2, 00:07:11.799 "num_base_bdevs_discovered": 2, 00:07:11.799 "num_base_bdevs_operational": 2, 00:07:11.799 "base_bdevs_list": [ 00:07:11.799 { 00:07:11.799 "name": "BaseBdev1", 00:07:11.799 "uuid": "a128fcfd-fe21-5333-be3a-fc6ad5e9fb41", 00:07:11.799 "is_configured": true, 00:07:11.799 "data_offset": 2048, 00:07:11.799 "data_size": 63488 00:07:11.799 }, 00:07:11.799 { 00:07:11.799 "name": "BaseBdev2", 00:07:11.799 "uuid": "4f014086-3c2b-51a5-95c0-caa4e6766383", 00:07:11.799 "is_configured": true, 00:07:11.799 "data_offset": 2048, 00:07:11.799 "data_size": 63488 00:07:11.799 } 00:07:11.799 ] 00:07:11.799 }' 00:07:11.799 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:11.799 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.059 [2024-09-29 16:45:33.558332] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:12.059 [2024-09-29 16:45:33.558367] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:12.059 [2024-09-29 16:45:33.560916] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:12.059 [2024-09-29 16:45:33.560961] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:12.059 [2024-09-29 16:45:33.560993] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:12.059 [2024-09-29 16:45:33.561003] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:12.059 { 00:07:12.059 "results": [ 00:07:12.059 { 00:07:12.059 "job": "raid_bdev1", 00:07:12.059 "core_mask": "0x1", 00:07:12.059 "workload": "randrw", 00:07:12.059 "percentage": 50, 00:07:12.059 "status": "finished", 00:07:12.059 "queue_depth": 1, 00:07:12.059 "io_size": 131072, 00:07:12.059 "runtime": 1.318443, 00:07:12.059 "iops": 17773.23706826916, 00:07:12.059 "mibps": 2221.654633533645, 00:07:12.059 "io_failed": 1, 00:07:12.059 "io_timeout": 0, 00:07:12.059 "avg_latency_us": 77.751873681841, 00:07:12.059 "min_latency_us": 24.482096069868994, 00:07:12.059 "max_latency_us": 1395.1441048034935 00:07:12.059 } 00:07:12.059 ], 00:07:12.059 "core_count": 1 00:07:12.059 } 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72490 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 72490 ']' 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 72490 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72490 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:12.059 killing process with pid 72490 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72490' 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 72490 00:07:12.059 [2024-09-29 16:45:33.607927] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:12.059 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 72490 00:07:12.059 [2024-09-29 16:45:33.623000] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.30mKPUYEYR 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.76 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.76 != \0\.\0\0 ]] 00:07:12.319 00:07:12.319 real 0m3.161s 00:07:12.319 user 0m3.972s 00:07:12.319 sys 0m0.486s 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:12.319 16:45:33 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.319 ************************************ 00:07:12.319 END TEST raid_read_error_test 00:07:12.319 ************************************ 00:07:12.319 16:45:33 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:07:12.319 16:45:33 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:12.319 16:45:33 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:12.319 16:45:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:12.319 ************************************ 00:07:12.319 START TEST raid_write_error_test 00:07:12.319 ************************************ 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 write 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Guiy3pY3DO 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72619 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72619 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 72619 ']' 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:12.319 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:12.319 16:45:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.579 [2024-09-29 16:45:34.032360] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:12.579 [2024-09-29 16:45:34.032468] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72619 ] 00:07:12.579 [2024-09-29 16:45:34.177377] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:12.579 [2024-09-29 16:45:34.224204] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:12.839 [2024-09-29 16:45:34.267411] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:12.839 [2024-09-29 16:45:34.267457] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 BaseBdev1_malloc 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 true 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 [2024-09-29 16:45:34.878152] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:13.411 [2024-09-29 16:45:34.878209] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:13.411 [2024-09-29 16:45:34.878239] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:13.411 [2024-09-29 16:45:34.878247] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:13.411 [2024-09-29 16:45:34.880359] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:13.411 [2024-09-29 16:45:34.880404] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:13.411 BaseBdev1 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 BaseBdev2_malloc 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 true 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 [2024-09-29 16:45:34.926452] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:13.411 [2024-09-29 16:45:34.926501] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:13.411 [2024-09-29 16:45:34.926521] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:13.411 [2024-09-29 16:45:34.926529] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:13.411 [2024-09-29 16:45:34.928560] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:13.411 [2024-09-29 16:45:34.928603] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:13.411 BaseBdev2 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 [2024-09-29 16:45:34.938503] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:13.411 [2024-09-29 16:45:34.940357] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:13.411 [2024-09-29 16:45:34.940529] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:13.411 [2024-09-29 16:45:34.940553] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:13.411 [2024-09-29 16:45:34.940835] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:13.411 [2024-09-29 16:45:34.940983] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:13.411 [2024-09-29 16:45:34.941004] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:13.411 [2024-09-29 16:45:34.941126] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:13.411 "name": "raid_bdev1", 00:07:13.411 "uuid": "82e7cdf2-aeff-4d2a-936a-b4c8c6e9dd07", 00:07:13.411 "strip_size_kb": 64, 00:07:13.411 "state": "online", 00:07:13.411 "raid_level": "raid0", 00:07:13.411 "superblock": true, 00:07:13.411 "num_base_bdevs": 2, 00:07:13.411 "num_base_bdevs_discovered": 2, 00:07:13.411 "num_base_bdevs_operational": 2, 00:07:13.411 "base_bdevs_list": [ 00:07:13.411 { 00:07:13.411 "name": "BaseBdev1", 00:07:13.411 "uuid": "a4fe0e7d-be71-55a8-a6eb-d14fb97739fe", 00:07:13.411 "is_configured": true, 00:07:13.411 "data_offset": 2048, 00:07:13.411 "data_size": 63488 00:07:13.411 }, 00:07:13.411 { 00:07:13.411 "name": "BaseBdev2", 00:07:13.411 "uuid": "0a7c524a-a930-55d7-896e-066c7faee291", 00:07:13.411 "is_configured": true, 00:07:13.411 "data_offset": 2048, 00:07:13.411 "data_size": 63488 00:07:13.411 } 00:07:13.411 ] 00:07:13.411 }' 00:07:13.411 16:45:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:13.412 16:45:34 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.981 16:45:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:13.981 16:45:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:13.981 [2024-09-29 16:45:35.457996] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:14.921 "name": "raid_bdev1", 00:07:14.921 "uuid": "82e7cdf2-aeff-4d2a-936a-b4c8c6e9dd07", 00:07:14.921 "strip_size_kb": 64, 00:07:14.921 "state": "online", 00:07:14.921 "raid_level": "raid0", 00:07:14.921 "superblock": true, 00:07:14.921 "num_base_bdevs": 2, 00:07:14.921 "num_base_bdevs_discovered": 2, 00:07:14.921 "num_base_bdevs_operational": 2, 00:07:14.921 "base_bdevs_list": [ 00:07:14.921 { 00:07:14.921 "name": "BaseBdev1", 00:07:14.921 "uuid": "a4fe0e7d-be71-55a8-a6eb-d14fb97739fe", 00:07:14.921 "is_configured": true, 00:07:14.921 "data_offset": 2048, 00:07:14.921 "data_size": 63488 00:07:14.921 }, 00:07:14.921 { 00:07:14.921 "name": "BaseBdev2", 00:07:14.921 "uuid": "0a7c524a-a930-55d7-896e-066c7faee291", 00:07:14.921 "is_configured": true, 00:07:14.921 "data_offset": 2048, 00:07:14.921 "data_size": 63488 00:07:14.921 } 00:07:14.921 ] 00:07:14.921 }' 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:14.921 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.182 [2024-09-29 16:45:36.807704] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:15.182 [2024-09-29 16:45:36.807799] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:15.182 [2024-09-29 16:45:36.810305] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:15.182 [2024-09-29 16:45:36.810382] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:15.182 [2024-09-29 16:45:36.810434] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:15.182 [2024-09-29 16:45:36.810473] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:15.182 { 00:07:15.182 "results": [ 00:07:15.182 { 00:07:15.182 "job": "raid_bdev1", 00:07:15.182 "core_mask": "0x1", 00:07:15.182 "workload": "randrw", 00:07:15.182 "percentage": 50, 00:07:15.182 "status": "finished", 00:07:15.182 "queue_depth": 1, 00:07:15.182 "io_size": 131072, 00:07:15.182 "runtime": 1.35061, 00:07:15.182 "iops": 17783.075795381345, 00:07:15.182 "mibps": 2222.884474422668, 00:07:15.182 "io_failed": 1, 00:07:15.182 "io_timeout": 0, 00:07:15.182 "avg_latency_us": 77.81077243979522, 00:07:15.182 "min_latency_us": 24.593886462882097, 00:07:15.182 "max_latency_us": 1495.3082969432314 00:07:15.182 } 00:07:15.182 ], 00:07:15.182 "core_count": 1 00:07:15.182 } 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72619 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 72619 ']' 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 72619 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:15.182 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72619 00:07:15.442 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:15.442 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:15.442 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72619' 00:07:15.442 killing process with pid 72619 00:07:15.442 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 72619 00:07:15.442 [2024-09-29 16:45:36.856163] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:15.442 16:45:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 72619 00:07:15.442 [2024-09-29 16:45:36.871169] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Guiy3pY3DO 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:15.442 00:07:15.442 real 0m3.177s 00:07:15.442 user 0m4.025s 00:07:15.442 sys 0m0.490s 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:15.442 ************************************ 00:07:15.442 END TEST raid_write_error_test 00:07:15.442 ************************************ 00:07:15.442 16:45:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.708 16:45:37 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:15.708 16:45:37 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:07:15.708 16:45:37 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:15.708 16:45:37 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:15.708 16:45:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:15.708 ************************************ 00:07:15.708 START TEST raid_state_function_test 00:07:15.708 ************************************ 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 false 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72746 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72746' 00:07:15.708 Process raid pid: 72746 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72746 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 72746 ']' 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:15.708 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:15.708 16:45:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.708 [2024-09-29 16:45:37.276301] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:15.708 [2024-09-29 16:45:37.276517] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:15.984 [2024-09-29 16:45:37.420296] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:15.984 [2024-09-29 16:45:37.467235] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:15.984 [2024-09-29 16:45:37.509912] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.984 [2024-09-29 16:45:37.510019] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.552 [2024-09-29 16:45:38.095801] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:16.552 [2024-09-29 16:45:38.095849] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:16.552 [2024-09-29 16:45:38.095861] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:16.552 [2024-09-29 16:45:38.095871] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:16.552 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:16.552 "name": "Existed_Raid", 00:07:16.552 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.552 "strip_size_kb": 64, 00:07:16.552 "state": "configuring", 00:07:16.552 "raid_level": "concat", 00:07:16.552 "superblock": false, 00:07:16.552 "num_base_bdevs": 2, 00:07:16.552 "num_base_bdevs_discovered": 0, 00:07:16.552 "num_base_bdevs_operational": 2, 00:07:16.552 "base_bdevs_list": [ 00:07:16.552 { 00:07:16.552 "name": "BaseBdev1", 00:07:16.552 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.553 "is_configured": false, 00:07:16.553 "data_offset": 0, 00:07:16.553 "data_size": 0 00:07:16.553 }, 00:07:16.553 { 00:07:16.553 "name": "BaseBdev2", 00:07:16.553 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:16.553 "is_configured": false, 00:07:16.553 "data_offset": 0, 00:07:16.553 "data_size": 0 00:07:16.553 } 00:07:16.553 ] 00:07:16.553 }' 00:07:16.553 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:16.553 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.122 [2024-09-29 16:45:38.538915] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:17.122 [2024-09-29 16:45:38.538993] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.122 [2024-09-29 16:45:38.550910] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:17.122 [2024-09-29 16:45:38.551002] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:17.122 [2024-09-29 16:45:38.551041] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:17.122 [2024-09-29 16:45:38.551064] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.122 [2024-09-29 16:45:38.571919] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:17.122 BaseBdev1 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.122 [ 00:07:17.122 { 00:07:17.122 "name": "BaseBdev1", 00:07:17.122 "aliases": [ 00:07:17.122 "0b538f2b-f6fe-44a4-b86c-50f210fd2b92" 00:07:17.122 ], 00:07:17.122 "product_name": "Malloc disk", 00:07:17.122 "block_size": 512, 00:07:17.122 "num_blocks": 65536, 00:07:17.122 "uuid": "0b538f2b-f6fe-44a4-b86c-50f210fd2b92", 00:07:17.122 "assigned_rate_limits": { 00:07:17.122 "rw_ios_per_sec": 0, 00:07:17.122 "rw_mbytes_per_sec": 0, 00:07:17.122 "r_mbytes_per_sec": 0, 00:07:17.122 "w_mbytes_per_sec": 0 00:07:17.122 }, 00:07:17.122 "claimed": true, 00:07:17.122 "claim_type": "exclusive_write", 00:07:17.122 "zoned": false, 00:07:17.122 "supported_io_types": { 00:07:17.122 "read": true, 00:07:17.122 "write": true, 00:07:17.122 "unmap": true, 00:07:17.122 "flush": true, 00:07:17.122 "reset": true, 00:07:17.122 "nvme_admin": false, 00:07:17.122 "nvme_io": false, 00:07:17.122 "nvme_io_md": false, 00:07:17.122 "write_zeroes": true, 00:07:17.122 "zcopy": true, 00:07:17.122 "get_zone_info": false, 00:07:17.122 "zone_management": false, 00:07:17.122 "zone_append": false, 00:07:17.122 "compare": false, 00:07:17.122 "compare_and_write": false, 00:07:17.122 "abort": true, 00:07:17.122 "seek_hole": false, 00:07:17.122 "seek_data": false, 00:07:17.122 "copy": true, 00:07:17.122 "nvme_iov_md": false 00:07:17.122 }, 00:07:17.122 "memory_domains": [ 00:07:17.122 { 00:07:17.122 "dma_device_id": "system", 00:07:17.122 "dma_device_type": 1 00:07:17.122 }, 00:07:17.122 { 00:07:17.122 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:17.122 "dma_device_type": 2 00:07:17.122 } 00:07:17.122 ], 00:07:17.122 "driver_specific": {} 00:07:17.122 } 00:07:17.122 ] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.122 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:17.122 "name": "Existed_Raid", 00:07:17.122 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.122 "strip_size_kb": 64, 00:07:17.122 "state": "configuring", 00:07:17.122 "raid_level": "concat", 00:07:17.122 "superblock": false, 00:07:17.122 "num_base_bdevs": 2, 00:07:17.122 "num_base_bdevs_discovered": 1, 00:07:17.122 "num_base_bdevs_operational": 2, 00:07:17.122 "base_bdevs_list": [ 00:07:17.122 { 00:07:17.122 "name": "BaseBdev1", 00:07:17.122 "uuid": "0b538f2b-f6fe-44a4-b86c-50f210fd2b92", 00:07:17.122 "is_configured": true, 00:07:17.122 "data_offset": 0, 00:07:17.122 "data_size": 65536 00:07:17.122 }, 00:07:17.122 { 00:07:17.122 "name": "BaseBdev2", 00:07:17.122 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.122 "is_configured": false, 00:07:17.122 "data_offset": 0, 00:07:17.122 "data_size": 0 00:07:17.122 } 00:07:17.122 ] 00:07:17.123 }' 00:07:17.123 16:45:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:17.123 16:45:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.382 [2024-09-29 16:45:39.027180] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:17.382 [2024-09-29 16:45:39.027232] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.382 [2024-09-29 16:45:39.039212] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:17.382 [2024-09-29 16:45:39.041029] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:17.382 [2024-09-29 16:45:39.041113] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.382 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.641 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.641 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:17.641 "name": "Existed_Raid", 00:07:17.641 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.641 "strip_size_kb": 64, 00:07:17.641 "state": "configuring", 00:07:17.641 "raid_level": "concat", 00:07:17.641 "superblock": false, 00:07:17.641 "num_base_bdevs": 2, 00:07:17.641 "num_base_bdevs_discovered": 1, 00:07:17.641 "num_base_bdevs_operational": 2, 00:07:17.641 "base_bdevs_list": [ 00:07:17.641 { 00:07:17.641 "name": "BaseBdev1", 00:07:17.641 "uuid": "0b538f2b-f6fe-44a4-b86c-50f210fd2b92", 00:07:17.641 "is_configured": true, 00:07:17.641 "data_offset": 0, 00:07:17.641 "data_size": 65536 00:07:17.641 }, 00:07:17.641 { 00:07:17.641 "name": "BaseBdev2", 00:07:17.641 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:17.641 "is_configured": false, 00:07:17.641 "data_offset": 0, 00:07:17.641 "data_size": 0 00:07:17.641 } 00:07:17.641 ] 00:07:17.641 }' 00:07:17.641 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:17.641 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.901 [2024-09-29 16:45:39.498475] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:17.901 [2024-09-29 16:45:39.498785] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:17.901 [2024-09-29 16:45:39.498959] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:17.901 [2024-09-29 16:45:39.500023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:17.901 [2024-09-29 16:45:39.500543] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:17.901 [2024-09-29 16:45:39.500683] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:17.901 [2024-09-29 16:45:39.501425] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:17.901 BaseBdev2 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.901 [ 00:07:17.901 { 00:07:17.901 "name": "BaseBdev2", 00:07:17.901 "aliases": [ 00:07:17.901 "80f4ee7d-6cb4-41b8-a3fb-36cf3314da69" 00:07:17.901 ], 00:07:17.901 "product_name": "Malloc disk", 00:07:17.901 "block_size": 512, 00:07:17.901 "num_blocks": 65536, 00:07:17.901 "uuid": "80f4ee7d-6cb4-41b8-a3fb-36cf3314da69", 00:07:17.901 "assigned_rate_limits": { 00:07:17.901 "rw_ios_per_sec": 0, 00:07:17.901 "rw_mbytes_per_sec": 0, 00:07:17.901 "r_mbytes_per_sec": 0, 00:07:17.901 "w_mbytes_per_sec": 0 00:07:17.901 }, 00:07:17.901 "claimed": true, 00:07:17.901 "claim_type": "exclusive_write", 00:07:17.901 "zoned": false, 00:07:17.901 "supported_io_types": { 00:07:17.901 "read": true, 00:07:17.901 "write": true, 00:07:17.901 "unmap": true, 00:07:17.901 "flush": true, 00:07:17.901 "reset": true, 00:07:17.901 "nvme_admin": false, 00:07:17.901 "nvme_io": false, 00:07:17.901 "nvme_io_md": false, 00:07:17.901 "write_zeroes": true, 00:07:17.901 "zcopy": true, 00:07:17.901 "get_zone_info": false, 00:07:17.901 "zone_management": false, 00:07:17.901 "zone_append": false, 00:07:17.901 "compare": false, 00:07:17.901 "compare_and_write": false, 00:07:17.901 "abort": true, 00:07:17.901 "seek_hole": false, 00:07:17.901 "seek_data": false, 00:07:17.901 "copy": true, 00:07:17.901 "nvme_iov_md": false 00:07:17.901 }, 00:07:17.901 "memory_domains": [ 00:07:17.901 { 00:07:17.901 "dma_device_id": "system", 00:07:17.901 "dma_device_type": 1 00:07:17.901 }, 00:07:17.901 { 00:07:17.901 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:17.901 "dma_device_type": 2 00:07:17.901 } 00:07:17.901 ], 00:07:17.901 "driver_specific": {} 00:07:17.901 } 00:07:17.901 ] 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:17.901 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.902 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:17.902 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.161 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.161 "name": "Existed_Raid", 00:07:18.161 "uuid": "7b1d6a3e-d1ab-47a1-ab28-9b0e85ec9b90", 00:07:18.161 "strip_size_kb": 64, 00:07:18.161 "state": "online", 00:07:18.161 "raid_level": "concat", 00:07:18.161 "superblock": false, 00:07:18.161 "num_base_bdevs": 2, 00:07:18.161 "num_base_bdevs_discovered": 2, 00:07:18.161 "num_base_bdevs_operational": 2, 00:07:18.161 "base_bdevs_list": [ 00:07:18.161 { 00:07:18.161 "name": "BaseBdev1", 00:07:18.161 "uuid": "0b538f2b-f6fe-44a4-b86c-50f210fd2b92", 00:07:18.161 "is_configured": true, 00:07:18.161 "data_offset": 0, 00:07:18.161 "data_size": 65536 00:07:18.161 }, 00:07:18.161 { 00:07:18.161 "name": "BaseBdev2", 00:07:18.161 "uuid": "80f4ee7d-6cb4-41b8-a3fb-36cf3314da69", 00:07:18.161 "is_configured": true, 00:07:18.161 "data_offset": 0, 00:07:18.161 "data_size": 65536 00:07:18.161 } 00:07:18.161 ] 00:07:18.161 }' 00:07:18.161 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.161 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:18.420 [2024-09-29 16:45:39.973876] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:18.420 16:45:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.420 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:18.420 "name": "Existed_Raid", 00:07:18.420 "aliases": [ 00:07:18.420 "7b1d6a3e-d1ab-47a1-ab28-9b0e85ec9b90" 00:07:18.420 ], 00:07:18.420 "product_name": "Raid Volume", 00:07:18.420 "block_size": 512, 00:07:18.420 "num_blocks": 131072, 00:07:18.420 "uuid": "7b1d6a3e-d1ab-47a1-ab28-9b0e85ec9b90", 00:07:18.420 "assigned_rate_limits": { 00:07:18.420 "rw_ios_per_sec": 0, 00:07:18.420 "rw_mbytes_per_sec": 0, 00:07:18.420 "r_mbytes_per_sec": 0, 00:07:18.420 "w_mbytes_per_sec": 0 00:07:18.420 }, 00:07:18.420 "claimed": false, 00:07:18.420 "zoned": false, 00:07:18.420 "supported_io_types": { 00:07:18.420 "read": true, 00:07:18.420 "write": true, 00:07:18.420 "unmap": true, 00:07:18.420 "flush": true, 00:07:18.420 "reset": true, 00:07:18.420 "nvme_admin": false, 00:07:18.420 "nvme_io": false, 00:07:18.420 "nvme_io_md": false, 00:07:18.420 "write_zeroes": true, 00:07:18.420 "zcopy": false, 00:07:18.420 "get_zone_info": false, 00:07:18.420 "zone_management": false, 00:07:18.420 "zone_append": false, 00:07:18.420 "compare": false, 00:07:18.420 "compare_and_write": false, 00:07:18.420 "abort": false, 00:07:18.420 "seek_hole": false, 00:07:18.420 "seek_data": false, 00:07:18.420 "copy": false, 00:07:18.420 "nvme_iov_md": false 00:07:18.420 }, 00:07:18.420 "memory_domains": [ 00:07:18.420 { 00:07:18.420 "dma_device_id": "system", 00:07:18.420 "dma_device_type": 1 00:07:18.420 }, 00:07:18.420 { 00:07:18.420 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.420 "dma_device_type": 2 00:07:18.420 }, 00:07:18.420 { 00:07:18.420 "dma_device_id": "system", 00:07:18.420 "dma_device_type": 1 00:07:18.420 }, 00:07:18.420 { 00:07:18.420 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:18.420 "dma_device_type": 2 00:07:18.420 } 00:07:18.420 ], 00:07:18.420 "driver_specific": { 00:07:18.420 "raid": { 00:07:18.421 "uuid": "7b1d6a3e-d1ab-47a1-ab28-9b0e85ec9b90", 00:07:18.421 "strip_size_kb": 64, 00:07:18.421 "state": "online", 00:07:18.421 "raid_level": "concat", 00:07:18.421 "superblock": false, 00:07:18.421 "num_base_bdevs": 2, 00:07:18.421 "num_base_bdevs_discovered": 2, 00:07:18.421 "num_base_bdevs_operational": 2, 00:07:18.421 "base_bdevs_list": [ 00:07:18.421 { 00:07:18.421 "name": "BaseBdev1", 00:07:18.421 "uuid": "0b538f2b-f6fe-44a4-b86c-50f210fd2b92", 00:07:18.421 "is_configured": true, 00:07:18.421 "data_offset": 0, 00:07:18.421 "data_size": 65536 00:07:18.421 }, 00:07:18.421 { 00:07:18.421 "name": "BaseBdev2", 00:07:18.421 "uuid": "80f4ee7d-6cb4-41b8-a3fb-36cf3314da69", 00:07:18.421 "is_configured": true, 00:07:18.421 "data_offset": 0, 00:07:18.421 "data_size": 65536 00:07:18.421 } 00:07:18.421 ] 00:07:18.421 } 00:07:18.421 } 00:07:18.421 }' 00:07:18.421 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:18.421 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:18.421 BaseBdev2' 00:07:18.421 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.681 [2024-09-29 16:45:40.225236] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:18.681 [2024-09-29 16:45:40.225262] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:18.681 [2024-09-29 16:45:40.225312] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:18.681 "name": "Existed_Raid", 00:07:18.681 "uuid": "7b1d6a3e-d1ab-47a1-ab28-9b0e85ec9b90", 00:07:18.681 "strip_size_kb": 64, 00:07:18.681 "state": "offline", 00:07:18.681 "raid_level": "concat", 00:07:18.681 "superblock": false, 00:07:18.681 "num_base_bdevs": 2, 00:07:18.681 "num_base_bdevs_discovered": 1, 00:07:18.681 "num_base_bdevs_operational": 1, 00:07:18.681 "base_bdevs_list": [ 00:07:18.681 { 00:07:18.681 "name": null, 00:07:18.681 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:18.681 "is_configured": false, 00:07:18.681 "data_offset": 0, 00:07:18.681 "data_size": 65536 00:07:18.681 }, 00:07:18.681 { 00:07:18.681 "name": "BaseBdev2", 00:07:18.681 "uuid": "80f4ee7d-6cb4-41b8-a3fb-36cf3314da69", 00:07:18.681 "is_configured": true, 00:07:18.681 "data_offset": 0, 00:07:18.681 "data_size": 65536 00:07:18.681 } 00:07:18.681 ] 00:07:18.681 }' 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:18.681 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.248 [2024-09-29 16:45:40.707555] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:19.248 [2024-09-29 16:45:40.707605] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.248 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72746 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 72746 ']' 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 72746 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72746 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72746' 00:07:19.249 killing process with pid 72746 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 72746 00:07:19.249 [2024-09-29 16:45:40.813586] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:19.249 16:45:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 72746 00:07:19.249 [2024-09-29 16:45:40.814549] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:19.509 00:07:19.509 real 0m3.869s 00:07:19.509 user 0m6.089s 00:07:19.509 sys 0m0.745s 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:19.509 ************************************ 00:07:19.509 END TEST raid_state_function_test 00:07:19.509 ************************************ 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.509 16:45:41 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:07:19.509 16:45:41 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:19.509 16:45:41 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:19.509 16:45:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:19.509 ************************************ 00:07:19.509 START TEST raid_state_function_test_sb 00:07:19.509 ************************************ 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 true 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72988 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72988' 00:07:19.509 Process raid pid: 72988 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72988 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 72988 ']' 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:19.509 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:19.509 16:45:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:19.769 [2024-09-29 16:45:41.214149] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:19.769 [2024-09-29 16:45:41.214359] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:19.769 [2024-09-29 16:45:41.339124] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:19.769 [2024-09-29 16:45:41.383854] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:19.769 [2024-09-29 16:45:41.425200] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:19.769 [2024-09-29 16:45:41.425242] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:20.708 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:20.708 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:20.708 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:20.708 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.709 [2024-09-29 16:45:42.034401] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:20.709 [2024-09-29 16:45:42.034532] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:20.709 [2024-09-29 16:45:42.034551] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:20.709 [2024-09-29 16:45:42.034561] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.709 "name": "Existed_Raid", 00:07:20.709 "uuid": "eec9f013-63d0-440c-a0d7-75b92fbd682c", 00:07:20.709 "strip_size_kb": 64, 00:07:20.709 "state": "configuring", 00:07:20.709 "raid_level": "concat", 00:07:20.709 "superblock": true, 00:07:20.709 "num_base_bdevs": 2, 00:07:20.709 "num_base_bdevs_discovered": 0, 00:07:20.709 "num_base_bdevs_operational": 2, 00:07:20.709 "base_bdevs_list": [ 00:07:20.709 { 00:07:20.709 "name": "BaseBdev1", 00:07:20.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.709 "is_configured": false, 00:07:20.709 "data_offset": 0, 00:07:20.709 "data_size": 0 00:07:20.709 }, 00:07:20.709 { 00:07:20.709 "name": "BaseBdev2", 00:07:20.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.709 "is_configured": false, 00:07:20.709 "data_offset": 0, 00:07:20.709 "data_size": 0 00:07:20.709 } 00:07:20.709 ] 00:07:20.709 }' 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.709 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.968 [2024-09-29 16:45:42.485530] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:20.968 [2024-09-29 16:45:42.485620] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.968 [2024-09-29 16:45:42.497526] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:20.968 [2024-09-29 16:45:42.497600] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:20.968 [2024-09-29 16:45:42.497635] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:20.968 [2024-09-29 16:45:42.497658] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.968 [2024-09-29 16:45:42.518520] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:20.968 BaseBdev1 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.968 [ 00:07:20.968 { 00:07:20.968 "name": "BaseBdev1", 00:07:20.968 "aliases": [ 00:07:20.968 "66d9c092-07a2-4823-99ce-ff2260c411ab" 00:07:20.968 ], 00:07:20.968 "product_name": "Malloc disk", 00:07:20.968 "block_size": 512, 00:07:20.968 "num_blocks": 65536, 00:07:20.968 "uuid": "66d9c092-07a2-4823-99ce-ff2260c411ab", 00:07:20.968 "assigned_rate_limits": { 00:07:20.968 "rw_ios_per_sec": 0, 00:07:20.968 "rw_mbytes_per_sec": 0, 00:07:20.968 "r_mbytes_per_sec": 0, 00:07:20.968 "w_mbytes_per_sec": 0 00:07:20.968 }, 00:07:20.968 "claimed": true, 00:07:20.968 "claim_type": "exclusive_write", 00:07:20.968 "zoned": false, 00:07:20.968 "supported_io_types": { 00:07:20.968 "read": true, 00:07:20.968 "write": true, 00:07:20.968 "unmap": true, 00:07:20.968 "flush": true, 00:07:20.968 "reset": true, 00:07:20.968 "nvme_admin": false, 00:07:20.968 "nvme_io": false, 00:07:20.968 "nvme_io_md": false, 00:07:20.968 "write_zeroes": true, 00:07:20.968 "zcopy": true, 00:07:20.968 "get_zone_info": false, 00:07:20.968 "zone_management": false, 00:07:20.968 "zone_append": false, 00:07:20.968 "compare": false, 00:07:20.968 "compare_and_write": false, 00:07:20.968 "abort": true, 00:07:20.968 "seek_hole": false, 00:07:20.968 "seek_data": false, 00:07:20.968 "copy": true, 00:07:20.968 "nvme_iov_md": false 00:07:20.968 }, 00:07:20.968 "memory_domains": [ 00:07:20.968 { 00:07:20.968 "dma_device_id": "system", 00:07:20.968 "dma_device_type": 1 00:07:20.968 }, 00:07:20.968 { 00:07:20.968 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:20.968 "dma_device_type": 2 00:07:20.968 } 00:07:20.968 ], 00:07:20.968 "driver_specific": {} 00:07:20.968 } 00:07:20.968 ] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.968 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:20.968 "name": "Existed_Raid", 00:07:20.968 "uuid": "71450080-d096-4658-b99c-c57a78377e22", 00:07:20.968 "strip_size_kb": 64, 00:07:20.968 "state": "configuring", 00:07:20.968 "raid_level": "concat", 00:07:20.968 "superblock": true, 00:07:20.968 "num_base_bdevs": 2, 00:07:20.968 "num_base_bdevs_discovered": 1, 00:07:20.968 "num_base_bdevs_operational": 2, 00:07:20.968 "base_bdevs_list": [ 00:07:20.968 { 00:07:20.968 "name": "BaseBdev1", 00:07:20.968 "uuid": "66d9c092-07a2-4823-99ce-ff2260c411ab", 00:07:20.968 "is_configured": true, 00:07:20.968 "data_offset": 2048, 00:07:20.968 "data_size": 63488 00:07:20.968 }, 00:07:20.968 { 00:07:20.968 "name": "BaseBdev2", 00:07:20.968 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:20.968 "is_configured": false, 00:07:20.968 "data_offset": 0, 00:07:20.968 "data_size": 0 00:07:20.969 } 00:07:20.969 ] 00:07:20.969 }' 00:07:20.969 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:20.969 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.537 [2024-09-29 16:45:42.981785] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:21.537 [2024-09-29 16:45:42.981828] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.537 [2024-09-29 16:45:42.993812] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:21.537 [2024-09-29 16:45:42.995664] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:21.537 [2024-09-29 16:45:42.995706] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:21.537 16:45:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:21.537 "name": "Existed_Raid", 00:07:21.537 "uuid": "3087b8e7-ba7b-4999-bd4a-e9ed3aa11d05", 00:07:21.537 "strip_size_kb": 64, 00:07:21.537 "state": "configuring", 00:07:21.537 "raid_level": "concat", 00:07:21.537 "superblock": true, 00:07:21.537 "num_base_bdevs": 2, 00:07:21.537 "num_base_bdevs_discovered": 1, 00:07:21.537 "num_base_bdevs_operational": 2, 00:07:21.537 "base_bdevs_list": [ 00:07:21.537 { 00:07:21.537 "name": "BaseBdev1", 00:07:21.537 "uuid": "66d9c092-07a2-4823-99ce-ff2260c411ab", 00:07:21.537 "is_configured": true, 00:07:21.537 "data_offset": 2048, 00:07:21.537 "data_size": 63488 00:07:21.537 }, 00:07:21.537 { 00:07:21.537 "name": "BaseBdev2", 00:07:21.537 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:21.537 "is_configured": false, 00:07:21.537 "data_offset": 0, 00:07:21.537 "data_size": 0 00:07:21.537 } 00:07:21.537 ] 00:07:21.537 }' 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:21.537 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:21.796 [2024-09-29 16:45:43.461618] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:21.796 [2024-09-29 16:45:43.461940] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:21.796 [2024-09-29 16:45:43.462005] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:21.796 [2024-09-29 16:45:43.462392] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:21.796 BaseBdev2 00:07:21.796 [2024-09-29 16:45:43.462610] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:21.796 [2024-09-29 16:45:43.462638] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:21.796 [2024-09-29 16:45:43.462837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:21.796 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.055 [ 00:07:22.055 { 00:07:22.055 "name": "BaseBdev2", 00:07:22.055 "aliases": [ 00:07:22.055 "c75580e7-0deb-46ec-a118-fe61c86534d7" 00:07:22.055 ], 00:07:22.055 "product_name": "Malloc disk", 00:07:22.055 "block_size": 512, 00:07:22.055 "num_blocks": 65536, 00:07:22.055 "uuid": "c75580e7-0deb-46ec-a118-fe61c86534d7", 00:07:22.055 "assigned_rate_limits": { 00:07:22.055 "rw_ios_per_sec": 0, 00:07:22.055 "rw_mbytes_per_sec": 0, 00:07:22.055 "r_mbytes_per_sec": 0, 00:07:22.055 "w_mbytes_per_sec": 0 00:07:22.055 }, 00:07:22.055 "claimed": true, 00:07:22.055 "claim_type": "exclusive_write", 00:07:22.055 "zoned": false, 00:07:22.055 "supported_io_types": { 00:07:22.055 "read": true, 00:07:22.055 "write": true, 00:07:22.055 "unmap": true, 00:07:22.055 "flush": true, 00:07:22.055 "reset": true, 00:07:22.055 "nvme_admin": false, 00:07:22.055 "nvme_io": false, 00:07:22.055 "nvme_io_md": false, 00:07:22.055 "write_zeroes": true, 00:07:22.055 "zcopy": true, 00:07:22.055 "get_zone_info": false, 00:07:22.055 "zone_management": false, 00:07:22.055 "zone_append": false, 00:07:22.055 "compare": false, 00:07:22.055 "compare_and_write": false, 00:07:22.055 "abort": true, 00:07:22.055 "seek_hole": false, 00:07:22.055 "seek_data": false, 00:07:22.055 "copy": true, 00:07:22.055 "nvme_iov_md": false 00:07:22.055 }, 00:07:22.055 "memory_domains": [ 00:07:22.055 { 00:07:22.055 "dma_device_id": "system", 00:07:22.055 "dma_device_type": 1 00:07:22.055 }, 00:07:22.055 { 00:07:22.055 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.055 "dma_device_type": 2 00:07:22.055 } 00:07:22.055 ], 00:07:22.055 "driver_specific": {} 00:07:22.055 } 00:07:22.055 ] 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:22.055 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.056 "name": "Existed_Raid", 00:07:22.056 "uuid": "3087b8e7-ba7b-4999-bd4a-e9ed3aa11d05", 00:07:22.056 "strip_size_kb": 64, 00:07:22.056 "state": "online", 00:07:22.056 "raid_level": "concat", 00:07:22.056 "superblock": true, 00:07:22.056 "num_base_bdevs": 2, 00:07:22.056 "num_base_bdevs_discovered": 2, 00:07:22.056 "num_base_bdevs_operational": 2, 00:07:22.056 "base_bdevs_list": [ 00:07:22.056 { 00:07:22.056 "name": "BaseBdev1", 00:07:22.056 "uuid": "66d9c092-07a2-4823-99ce-ff2260c411ab", 00:07:22.056 "is_configured": true, 00:07:22.056 "data_offset": 2048, 00:07:22.056 "data_size": 63488 00:07:22.056 }, 00:07:22.056 { 00:07:22.056 "name": "BaseBdev2", 00:07:22.056 "uuid": "c75580e7-0deb-46ec-a118-fe61c86534d7", 00:07:22.056 "is_configured": true, 00:07:22.056 "data_offset": 2048, 00:07:22.056 "data_size": 63488 00:07:22.056 } 00:07:22.056 ] 00:07:22.056 }' 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.056 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.316 [2024-09-29 16:45:43.937097] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:22.316 "name": "Existed_Raid", 00:07:22.316 "aliases": [ 00:07:22.316 "3087b8e7-ba7b-4999-bd4a-e9ed3aa11d05" 00:07:22.316 ], 00:07:22.316 "product_name": "Raid Volume", 00:07:22.316 "block_size": 512, 00:07:22.316 "num_blocks": 126976, 00:07:22.316 "uuid": "3087b8e7-ba7b-4999-bd4a-e9ed3aa11d05", 00:07:22.316 "assigned_rate_limits": { 00:07:22.316 "rw_ios_per_sec": 0, 00:07:22.316 "rw_mbytes_per_sec": 0, 00:07:22.316 "r_mbytes_per_sec": 0, 00:07:22.316 "w_mbytes_per_sec": 0 00:07:22.316 }, 00:07:22.316 "claimed": false, 00:07:22.316 "zoned": false, 00:07:22.316 "supported_io_types": { 00:07:22.316 "read": true, 00:07:22.316 "write": true, 00:07:22.316 "unmap": true, 00:07:22.316 "flush": true, 00:07:22.316 "reset": true, 00:07:22.316 "nvme_admin": false, 00:07:22.316 "nvme_io": false, 00:07:22.316 "nvme_io_md": false, 00:07:22.316 "write_zeroes": true, 00:07:22.316 "zcopy": false, 00:07:22.316 "get_zone_info": false, 00:07:22.316 "zone_management": false, 00:07:22.316 "zone_append": false, 00:07:22.316 "compare": false, 00:07:22.316 "compare_and_write": false, 00:07:22.316 "abort": false, 00:07:22.316 "seek_hole": false, 00:07:22.316 "seek_data": false, 00:07:22.316 "copy": false, 00:07:22.316 "nvme_iov_md": false 00:07:22.316 }, 00:07:22.316 "memory_domains": [ 00:07:22.316 { 00:07:22.316 "dma_device_id": "system", 00:07:22.316 "dma_device_type": 1 00:07:22.316 }, 00:07:22.316 { 00:07:22.316 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.316 "dma_device_type": 2 00:07:22.316 }, 00:07:22.316 { 00:07:22.316 "dma_device_id": "system", 00:07:22.316 "dma_device_type": 1 00:07:22.316 }, 00:07:22.316 { 00:07:22.316 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:22.316 "dma_device_type": 2 00:07:22.316 } 00:07:22.316 ], 00:07:22.316 "driver_specific": { 00:07:22.316 "raid": { 00:07:22.316 "uuid": "3087b8e7-ba7b-4999-bd4a-e9ed3aa11d05", 00:07:22.316 "strip_size_kb": 64, 00:07:22.316 "state": "online", 00:07:22.316 "raid_level": "concat", 00:07:22.316 "superblock": true, 00:07:22.316 "num_base_bdevs": 2, 00:07:22.316 "num_base_bdevs_discovered": 2, 00:07:22.316 "num_base_bdevs_operational": 2, 00:07:22.316 "base_bdevs_list": [ 00:07:22.316 { 00:07:22.316 "name": "BaseBdev1", 00:07:22.316 "uuid": "66d9c092-07a2-4823-99ce-ff2260c411ab", 00:07:22.316 "is_configured": true, 00:07:22.316 "data_offset": 2048, 00:07:22.316 "data_size": 63488 00:07:22.316 }, 00:07:22.316 { 00:07:22.316 "name": "BaseBdev2", 00:07:22.316 "uuid": "c75580e7-0deb-46ec-a118-fe61c86534d7", 00:07:22.316 "is_configured": true, 00:07:22.316 "data_offset": 2048, 00:07:22.316 "data_size": 63488 00:07:22.316 } 00:07:22.316 ] 00:07:22.316 } 00:07:22.316 } 00:07:22.316 }' 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:22.316 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:22.316 BaseBdev2' 00:07:22.576 16:45:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.576 [2024-09-29 16:45:44.104610] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:22.576 [2024-09-29 16:45:44.104679] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:22.576 [2024-09-29 16:45:44.104749] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:22.576 "name": "Existed_Raid", 00:07:22.576 "uuid": "3087b8e7-ba7b-4999-bd4a-e9ed3aa11d05", 00:07:22.576 "strip_size_kb": 64, 00:07:22.576 "state": "offline", 00:07:22.576 "raid_level": "concat", 00:07:22.576 "superblock": true, 00:07:22.576 "num_base_bdevs": 2, 00:07:22.576 "num_base_bdevs_discovered": 1, 00:07:22.576 "num_base_bdevs_operational": 1, 00:07:22.576 "base_bdevs_list": [ 00:07:22.576 { 00:07:22.576 "name": null, 00:07:22.576 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:22.576 "is_configured": false, 00:07:22.576 "data_offset": 0, 00:07:22.576 "data_size": 63488 00:07:22.576 }, 00:07:22.576 { 00:07:22.576 "name": "BaseBdev2", 00:07:22.576 "uuid": "c75580e7-0deb-46ec-a118-fe61c86534d7", 00:07:22.576 "is_configured": true, 00:07:22.576 "data_offset": 2048, 00:07:22.576 "data_size": 63488 00:07:22.576 } 00:07:22.576 ] 00:07:22.576 }' 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:22.576 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.144 [2024-09-29 16:45:44.630989] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:23.144 [2024-09-29 16:45:44.631079] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72988 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 72988 ']' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 72988 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72988 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:23.144 killing process with pid 72988 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72988' 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 72988 00:07:23.144 [2024-09-29 16:45:44.734452] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:23.144 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 72988 00:07:23.144 [2024-09-29 16:45:44.735436] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:23.405 16:45:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:23.405 00:07:23.405 real 0m3.849s 00:07:23.405 user 0m6.044s 00:07:23.405 sys 0m0.771s 00:07:23.405 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:23.405 16:45:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:23.405 ************************************ 00:07:23.405 END TEST raid_state_function_test_sb 00:07:23.405 ************************************ 00:07:23.405 16:45:45 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:07:23.405 16:45:45 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:23.405 16:45:45 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:23.405 16:45:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:23.405 ************************************ 00:07:23.405 START TEST raid_superblock_test 00:07:23.405 ************************************ 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 2 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73229 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73229 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 73229 ']' 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:23.405 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:23.405 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:23.664 [2024-09-29 16:45:45.130079] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:23.664 [2024-09-29 16:45:45.130285] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73229 ] 00:07:23.664 [2024-09-29 16:45:45.275247] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:23.664 [2024-09-29 16:45:45.320644] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:23.922 [2024-09-29 16:45:45.362542] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:23.922 [2024-09-29 16:45:45.362661] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.490 malloc1 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.490 [2024-09-29 16:45:45.960290] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:24.490 [2024-09-29 16:45:45.960353] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:24.490 [2024-09-29 16:45:45.960375] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:24.490 [2024-09-29 16:45:45.960388] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:24.490 [2024-09-29 16:45:45.962452] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:24.490 [2024-09-29 16:45:45.962542] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:24.490 pt1 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:24.490 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.491 malloc2 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.491 16:45:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.491 [2024-09-29 16:45:46.004571] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:24.491 [2024-09-29 16:45:46.004769] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:24.491 [2024-09-29 16:45:46.004847] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:24.491 [2024-09-29 16:45:46.004926] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:24.491 [2024-09-29 16:45:46.009487] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:24.491 [2024-09-29 16:45:46.009637] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:24.491 pt2 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.491 [2024-09-29 16:45:46.017946] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:24.491 [2024-09-29 16:45:46.020428] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:24.491 [2024-09-29 16:45:46.020622] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:24.491 [2024-09-29 16:45:46.020689] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:24.491 [2024-09-29 16:45:46.021034] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:24.491 [2024-09-29 16:45:46.021225] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:24.491 [2024-09-29 16:45:46.021273] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:24.491 [2024-09-29 16:45:46.021487] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:24.491 "name": "raid_bdev1", 00:07:24.491 "uuid": "533214ee-399e-458a-ac40-cf8c88f483e8", 00:07:24.491 "strip_size_kb": 64, 00:07:24.491 "state": "online", 00:07:24.491 "raid_level": "concat", 00:07:24.491 "superblock": true, 00:07:24.491 "num_base_bdevs": 2, 00:07:24.491 "num_base_bdevs_discovered": 2, 00:07:24.491 "num_base_bdevs_operational": 2, 00:07:24.491 "base_bdevs_list": [ 00:07:24.491 { 00:07:24.491 "name": "pt1", 00:07:24.491 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:24.491 "is_configured": true, 00:07:24.491 "data_offset": 2048, 00:07:24.491 "data_size": 63488 00:07:24.491 }, 00:07:24.491 { 00:07:24.491 "name": "pt2", 00:07:24.491 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:24.491 "is_configured": true, 00:07:24.491 "data_offset": 2048, 00:07:24.491 "data_size": 63488 00:07:24.491 } 00:07:24.491 ] 00:07:24.491 }' 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:24.491 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.059 [2024-09-29 16:45:46.433384] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:25.059 "name": "raid_bdev1", 00:07:25.059 "aliases": [ 00:07:25.059 "533214ee-399e-458a-ac40-cf8c88f483e8" 00:07:25.059 ], 00:07:25.059 "product_name": "Raid Volume", 00:07:25.059 "block_size": 512, 00:07:25.059 "num_blocks": 126976, 00:07:25.059 "uuid": "533214ee-399e-458a-ac40-cf8c88f483e8", 00:07:25.059 "assigned_rate_limits": { 00:07:25.059 "rw_ios_per_sec": 0, 00:07:25.059 "rw_mbytes_per_sec": 0, 00:07:25.059 "r_mbytes_per_sec": 0, 00:07:25.059 "w_mbytes_per_sec": 0 00:07:25.059 }, 00:07:25.059 "claimed": false, 00:07:25.059 "zoned": false, 00:07:25.059 "supported_io_types": { 00:07:25.059 "read": true, 00:07:25.059 "write": true, 00:07:25.059 "unmap": true, 00:07:25.059 "flush": true, 00:07:25.059 "reset": true, 00:07:25.059 "nvme_admin": false, 00:07:25.059 "nvme_io": false, 00:07:25.059 "nvme_io_md": false, 00:07:25.059 "write_zeroes": true, 00:07:25.059 "zcopy": false, 00:07:25.059 "get_zone_info": false, 00:07:25.059 "zone_management": false, 00:07:25.059 "zone_append": false, 00:07:25.059 "compare": false, 00:07:25.059 "compare_and_write": false, 00:07:25.059 "abort": false, 00:07:25.059 "seek_hole": false, 00:07:25.059 "seek_data": false, 00:07:25.059 "copy": false, 00:07:25.059 "nvme_iov_md": false 00:07:25.059 }, 00:07:25.059 "memory_domains": [ 00:07:25.059 { 00:07:25.059 "dma_device_id": "system", 00:07:25.059 "dma_device_type": 1 00:07:25.059 }, 00:07:25.059 { 00:07:25.059 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:25.059 "dma_device_type": 2 00:07:25.059 }, 00:07:25.059 { 00:07:25.059 "dma_device_id": "system", 00:07:25.059 "dma_device_type": 1 00:07:25.059 }, 00:07:25.059 { 00:07:25.059 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:25.059 "dma_device_type": 2 00:07:25.059 } 00:07:25.059 ], 00:07:25.059 "driver_specific": { 00:07:25.059 "raid": { 00:07:25.059 "uuid": "533214ee-399e-458a-ac40-cf8c88f483e8", 00:07:25.059 "strip_size_kb": 64, 00:07:25.059 "state": "online", 00:07:25.059 "raid_level": "concat", 00:07:25.059 "superblock": true, 00:07:25.059 "num_base_bdevs": 2, 00:07:25.059 "num_base_bdevs_discovered": 2, 00:07:25.059 "num_base_bdevs_operational": 2, 00:07:25.059 "base_bdevs_list": [ 00:07:25.059 { 00:07:25.059 "name": "pt1", 00:07:25.059 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:25.059 "is_configured": true, 00:07:25.059 "data_offset": 2048, 00:07:25.059 "data_size": 63488 00:07:25.059 }, 00:07:25.059 { 00:07:25.059 "name": "pt2", 00:07:25.059 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:25.059 "is_configured": true, 00:07:25.059 "data_offset": 2048, 00:07:25.059 "data_size": 63488 00:07:25.059 } 00:07:25.059 ] 00:07:25.059 } 00:07:25.059 } 00:07:25.059 }' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:25.059 pt2' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:25.059 [2024-09-29 16:45:46.628978] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=533214ee-399e-458a-ac40-cf8c88f483e8 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 533214ee-399e-458a-ac40-cf8c88f483e8 ']' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.059 [2024-09-29 16:45:46.672671] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:25.059 [2024-09-29 16:45:46.672699] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:25.059 [2024-09-29 16:45:46.672785] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:25.059 [2024-09-29 16:45:46.672832] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:25.059 [2024-09-29 16:45:46.672843] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.059 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.318 [2024-09-29 16:45:46.804482] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:25.318 [2024-09-29 16:45:46.806300] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:25.318 [2024-09-29 16:45:46.806359] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:25.318 [2024-09-29 16:45:46.806411] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:25.318 [2024-09-29 16:45:46.806429] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:25.318 [2024-09-29 16:45:46.806438] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:25.318 request: 00:07:25.318 { 00:07:25.318 "name": "raid_bdev1", 00:07:25.318 "raid_level": "concat", 00:07:25.318 "base_bdevs": [ 00:07:25.318 "malloc1", 00:07:25.318 "malloc2" 00:07:25.318 ], 00:07:25.318 "strip_size_kb": 64, 00:07:25.318 "superblock": false, 00:07:25.318 "method": "bdev_raid_create", 00:07:25.318 "req_id": 1 00:07:25.318 } 00:07:25.318 Got JSON-RPC error response 00:07:25.318 response: 00:07:25.318 { 00:07:25.318 "code": -17, 00:07:25.318 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:25.318 } 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.318 [2024-09-29 16:45:46.868319] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:25.318 [2024-09-29 16:45:46.868411] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.318 [2024-09-29 16:45:46.868451] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:25.318 [2024-09-29 16:45:46.868479] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.318 [2024-09-29 16:45:46.870568] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.318 [2024-09-29 16:45:46.870651] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:25.318 [2024-09-29 16:45:46.870748] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:25.318 [2024-09-29 16:45:46.870814] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:25.318 pt1 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.318 "name": "raid_bdev1", 00:07:25.318 "uuid": "533214ee-399e-458a-ac40-cf8c88f483e8", 00:07:25.318 "strip_size_kb": 64, 00:07:25.318 "state": "configuring", 00:07:25.318 "raid_level": "concat", 00:07:25.318 "superblock": true, 00:07:25.318 "num_base_bdevs": 2, 00:07:25.318 "num_base_bdevs_discovered": 1, 00:07:25.318 "num_base_bdevs_operational": 2, 00:07:25.318 "base_bdevs_list": [ 00:07:25.318 { 00:07:25.318 "name": "pt1", 00:07:25.318 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:25.318 "is_configured": true, 00:07:25.318 "data_offset": 2048, 00:07:25.318 "data_size": 63488 00:07:25.318 }, 00:07:25.318 { 00:07:25.318 "name": null, 00:07:25.318 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:25.318 "is_configured": false, 00:07:25.318 "data_offset": 2048, 00:07:25.318 "data_size": 63488 00:07:25.318 } 00:07:25.318 ] 00:07:25.318 }' 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.318 16:45:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.887 [2024-09-29 16:45:47.287658] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:25.887 [2024-09-29 16:45:47.287762] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:25.887 [2024-09-29 16:45:47.287802] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:25.887 [2024-09-29 16:45:47.287830] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:25.887 [2024-09-29 16:45:47.288274] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:25.887 [2024-09-29 16:45:47.288350] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:25.887 [2024-09-29 16:45:47.288466] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:25.887 [2024-09-29 16:45:47.288515] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:25.887 [2024-09-29 16:45:47.288618] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:25.887 [2024-09-29 16:45:47.288628] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:25.887 [2024-09-29 16:45:47.288881] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:25.887 [2024-09-29 16:45:47.288983] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:25.887 [2024-09-29 16:45:47.288995] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:25.887 [2024-09-29 16:45:47.289092] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:25.887 pt2 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:25.887 "name": "raid_bdev1", 00:07:25.887 "uuid": "533214ee-399e-458a-ac40-cf8c88f483e8", 00:07:25.887 "strip_size_kb": 64, 00:07:25.887 "state": "online", 00:07:25.887 "raid_level": "concat", 00:07:25.887 "superblock": true, 00:07:25.887 "num_base_bdevs": 2, 00:07:25.887 "num_base_bdevs_discovered": 2, 00:07:25.887 "num_base_bdevs_operational": 2, 00:07:25.887 "base_bdevs_list": [ 00:07:25.887 { 00:07:25.887 "name": "pt1", 00:07:25.887 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:25.887 "is_configured": true, 00:07:25.887 "data_offset": 2048, 00:07:25.887 "data_size": 63488 00:07:25.887 }, 00:07:25.887 { 00:07:25.887 "name": "pt2", 00:07:25.887 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:25.887 "is_configured": true, 00:07:25.887 "data_offset": 2048, 00:07:25.887 "data_size": 63488 00:07:25.887 } 00:07:25.887 ] 00:07:25.887 }' 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:25.887 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.148 [2024-09-29 16:45:47.679269] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:26.148 "name": "raid_bdev1", 00:07:26.148 "aliases": [ 00:07:26.148 "533214ee-399e-458a-ac40-cf8c88f483e8" 00:07:26.148 ], 00:07:26.148 "product_name": "Raid Volume", 00:07:26.148 "block_size": 512, 00:07:26.148 "num_blocks": 126976, 00:07:26.148 "uuid": "533214ee-399e-458a-ac40-cf8c88f483e8", 00:07:26.148 "assigned_rate_limits": { 00:07:26.148 "rw_ios_per_sec": 0, 00:07:26.148 "rw_mbytes_per_sec": 0, 00:07:26.148 "r_mbytes_per_sec": 0, 00:07:26.148 "w_mbytes_per_sec": 0 00:07:26.148 }, 00:07:26.148 "claimed": false, 00:07:26.148 "zoned": false, 00:07:26.148 "supported_io_types": { 00:07:26.148 "read": true, 00:07:26.148 "write": true, 00:07:26.148 "unmap": true, 00:07:26.148 "flush": true, 00:07:26.148 "reset": true, 00:07:26.148 "nvme_admin": false, 00:07:26.148 "nvme_io": false, 00:07:26.148 "nvme_io_md": false, 00:07:26.148 "write_zeroes": true, 00:07:26.148 "zcopy": false, 00:07:26.148 "get_zone_info": false, 00:07:26.148 "zone_management": false, 00:07:26.148 "zone_append": false, 00:07:26.148 "compare": false, 00:07:26.148 "compare_and_write": false, 00:07:26.148 "abort": false, 00:07:26.148 "seek_hole": false, 00:07:26.148 "seek_data": false, 00:07:26.148 "copy": false, 00:07:26.148 "nvme_iov_md": false 00:07:26.148 }, 00:07:26.148 "memory_domains": [ 00:07:26.148 { 00:07:26.148 "dma_device_id": "system", 00:07:26.148 "dma_device_type": 1 00:07:26.148 }, 00:07:26.148 { 00:07:26.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.148 "dma_device_type": 2 00:07:26.148 }, 00:07:26.148 { 00:07:26.148 "dma_device_id": "system", 00:07:26.148 "dma_device_type": 1 00:07:26.148 }, 00:07:26.148 { 00:07:26.148 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:26.148 "dma_device_type": 2 00:07:26.148 } 00:07:26.148 ], 00:07:26.148 "driver_specific": { 00:07:26.148 "raid": { 00:07:26.148 "uuid": "533214ee-399e-458a-ac40-cf8c88f483e8", 00:07:26.148 "strip_size_kb": 64, 00:07:26.148 "state": "online", 00:07:26.148 "raid_level": "concat", 00:07:26.148 "superblock": true, 00:07:26.148 "num_base_bdevs": 2, 00:07:26.148 "num_base_bdevs_discovered": 2, 00:07:26.148 "num_base_bdevs_operational": 2, 00:07:26.148 "base_bdevs_list": [ 00:07:26.148 { 00:07:26.148 "name": "pt1", 00:07:26.148 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:26.148 "is_configured": true, 00:07:26.148 "data_offset": 2048, 00:07:26.148 "data_size": 63488 00:07:26.148 }, 00:07:26.148 { 00:07:26.148 "name": "pt2", 00:07:26.148 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:26.148 "is_configured": true, 00:07:26.148 "data_offset": 2048, 00:07:26.148 "data_size": 63488 00:07:26.148 } 00:07:26.148 ] 00:07:26.148 } 00:07:26.148 } 00:07:26.148 }' 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:26.148 pt2' 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.148 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.408 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.409 [2024-09-29 16:45:47.894855] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 533214ee-399e-458a-ac40-cf8c88f483e8 '!=' 533214ee-399e-458a-ac40-cf8c88f483e8 ']' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73229 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 73229 ']' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 73229 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73229 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73229' 00:07:26.409 killing process with pid 73229 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 73229 00:07:26.409 [2024-09-29 16:45:47.978939] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:26.409 [2024-09-29 16:45:47.979077] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.409 16:45:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 73229 00:07:26.409 [2024-09-29 16:45:47.979162] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.409 [2024-09-29 16:45:47.979182] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:26.409 [2024-09-29 16:45:48.001370] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:26.669 16:45:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:26.669 00:07:26.669 real 0m3.199s 00:07:26.669 user 0m4.865s 00:07:26.669 sys 0m0.701s 00:07:26.669 16:45:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:26.669 ************************************ 00:07:26.669 END TEST raid_superblock_test 00:07:26.669 ************************************ 00:07:26.669 16:45:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.669 16:45:48 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:07:26.669 16:45:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:26.669 16:45:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:26.669 16:45:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:26.669 ************************************ 00:07:26.669 START TEST raid_read_error_test 00:07:26.669 ************************************ 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 read 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.rhEtJUVRIC 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73424 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73424 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 73424 ']' 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:26.669 16:45:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:26.670 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:26.670 16:45:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:26.670 16:45:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.929 [2024-09-29 16:45:48.412338] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:26.929 [2024-09-29 16:45:48.412556] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73424 ] 00:07:26.929 [2024-09-29 16:45:48.551264] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:26.929 [2024-09-29 16:45:48.596015] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:27.188 [2024-09-29 16:45:48.637798] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.188 [2024-09-29 16:45:48.637832] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.756 BaseBdev1_malloc 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.756 true 00:07:27.756 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.757 [2024-09-29 16:45:49.268034] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:27.757 [2024-09-29 16:45:49.268091] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.757 [2024-09-29 16:45:49.268112] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:27.757 [2024-09-29 16:45:49.268121] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.757 [2024-09-29 16:45:49.270216] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.757 [2024-09-29 16:45:49.270253] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:27.757 BaseBdev1 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.757 BaseBdev2_malloc 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.757 true 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.757 [2024-09-29 16:45:49.324279] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:27.757 [2024-09-29 16:45:49.324359] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:27.757 [2024-09-29 16:45:49.324390] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:27.757 [2024-09-29 16:45:49.324402] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:27.757 [2024-09-29 16:45:49.327495] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:27.757 [2024-09-29 16:45:49.327609] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:27.757 BaseBdev2 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.757 [2024-09-29 16:45:49.336503] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:27.757 [2024-09-29 16:45:49.338529] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:27.757 [2024-09-29 16:45:49.338807] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:27.757 [2024-09-29 16:45:49.338827] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:27.757 [2024-09-29 16:45:49.339119] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:27.757 [2024-09-29 16:45:49.339265] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:27.757 [2024-09-29 16:45:49.339280] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:27.757 [2024-09-29 16:45:49.339419] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:27.757 "name": "raid_bdev1", 00:07:27.757 "uuid": "62868069-7b03-4c42-a573-d57e06a6dc75", 00:07:27.757 "strip_size_kb": 64, 00:07:27.757 "state": "online", 00:07:27.757 "raid_level": "concat", 00:07:27.757 "superblock": true, 00:07:27.757 "num_base_bdevs": 2, 00:07:27.757 "num_base_bdevs_discovered": 2, 00:07:27.757 "num_base_bdevs_operational": 2, 00:07:27.757 "base_bdevs_list": [ 00:07:27.757 { 00:07:27.757 "name": "BaseBdev1", 00:07:27.757 "uuid": "a94fe049-ed98-51b8-8d30-ef7b57ffc2f5", 00:07:27.757 "is_configured": true, 00:07:27.757 "data_offset": 2048, 00:07:27.757 "data_size": 63488 00:07:27.757 }, 00:07:27.757 { 00:07:27.757 "name": "BaseBdev2", 00:07:27.757 "uuid": "4de86c46-e0bb-5044-84b1-b35fd58f445d", 00:07:27.757 "is_configured": true, 00:07:27.757 "data_offset": 2048, 00:07:27.757 "data_size": 63488 00:07:27.757 } 00:07:27.757 ] 00:07:27.757 }' 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:27.757 16:45:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.411 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:28.411 16:45:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:28.411 [2024-09-29 16:45:49.863945] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.350 "name": "raid_bdev1", 00:07:29.350 "uuid": "62868069-7b03-4c42-a573-d57e06a6dc75", 00:07:29.350 "strip_size_kb": 64, 00:07:29.350 "state": "online", 00:07:29.350 "raid_level": "concat", 00:07:29.350 "superblock": true, 00:07:29.350 "num_base_bdevs": 2, 00:07:29.350 "num_base_bdevs_discovered": 2, 00:07:29.350 "num_base_bdevs_operational": 2, 00:07:29.350 "base_bdevs_list": [ 00:07:29.350 { 00:07:29.350 "name": "BaseBdev1", 00:07:29.350 "uuid": "a94fe049-ed98-51b8-8d30-ef7b57ffc2f5", 00:07:29.350 "is_configured": true, 00:07:29.350 "data_offset": 2048, 00:07:29.350 "data_size": 63488 00:07:29.350 }, 00:07:29.350 { 00:07:29.350 "name": "BaseBdev2", 00:07:29.350 "uuid": "4de86c46-e0bb-5044-84b1-b35fd58f445d", 00:07:29.350 "is_configured": true, 00:07:29.350 "data_offset": 2048, 00:07:29.350 "data_size": 63488 00:07:29.350 } 00:07:29.350 ] 00:07:29.350 }' 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.350 16:45:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.609 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:29.609 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.609 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.609 [2024-09-29 16:45:51.243515] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:29.609 [2024-09-29 16:45:51.243601] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:29.610 [2024-09-29 16:45:51.246177] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:29.610 [2024-09-29 16:45:51.246258] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:29.610 [2024-09-29 16:45:51.246309] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:29.610 [2024-09-29 16:45:51.246348] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:29.610 { 00:07:29.610 "results": [ 00:07:29.610 { 00:07:29.610 "job": "raid_bdev1", 00:07:29.610 "core_mask": "0x1", 00:07:29.610 "workload": "randrw", 00:07:29.610 "percentage": 50, 00:07:29.610 "status": "finished", 00:07:29.610 "queue_depth": 1, 00:07:29.610 "io_size": 131072, 00:07:29.610 "runtime": 1.38057, 00:07:29.610 "iops": 17875.225450357462, 00:07:29.610 "mibps": 2234.403181294683, 00:07:29.610 "io_failed": 1, 00:07:29.610 "io_timeout": 0, 00:07:29.610 "avg_latency_us": 77.25590177884031, 00:07:29.610 "min_latency_us": 24.258515283842794, 00:07:29.610 "max_latency_us": 1423.7624454148472 00:07:29.610 } 00:07:29.610 ], 00:07:29.610 "core_count": 1 00:07:29.610 } 00:07:29.610 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.610 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73424 00:07:29.610 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 73424 ']' 00:07:29.610 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 73424 00:07:29.610 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:07:29.610 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:29.610 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73424 00:07:29.870 killing process with pid 73424 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73424' 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 73424 00:07:29.870 [2024-09-29 16:45:51.293066] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 73424 00:07:29.870 [2024-09-29 16:45:51.308278] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.rhEtJUVRIC 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:29.870 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:30.130 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:07:30.130 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:30.130 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:30.130 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:30.130 16:45:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:07:30.130 00:07:30.130 real 0m3.229s 00:07:30.130 user 0m4.085s 00:07:30.130 sys 0m0.501s 00:07:30.130 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:30.130 16:45:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.130 ************************************ 00:07:30.130 END TEST raid_read_error_test 00:07:30.130 ************************************ 00:07:30.130 16:45:51 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:07:30.130 16:45:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:30.130 16:45:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:30.130 16:45:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:30.130 ************************************ 00:07:30.130 START TEST raid_write_error_test 00:07:30.130 ************************************ 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 write 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.bwaLzuO7Ki 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73553 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73553 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 73553 ']' 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:30.130 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:30.130 16:45:51 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.130 [2024-09-29 16:45:51.712879] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:30.130 [2024-09-29 16:45:51.713079] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73553 ] 00:07:30.390 [2024-09-29 16:45:51.856591] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:30.390 [2024-09-29 16:45:51.901939] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:30.390 [2024-09-29 16:45:51.943480] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:30.390 [2024-09-29 16:45:51.943604] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.959 BaseBdev1_malloc 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.959 true 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.959 [2024-09-29 16:45:52.565244] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:30.959 [2024-09-29 16:45:52.565321] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:30.959 [2024-09-29 16:45:52.565342] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:30.959 [2024-09-29 16:45:52.565350] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:30.959 [2024-09-29 16:45:52.567448] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:30.959 [2024-09-29 16:45:52.567534] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:30.959 BaseBdev1 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.959 BaseBdev2_malloc 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.959 true 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.959 [2024-09-29 16:45:52.623325] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:30.959 [2024-09-29 16:45:52.623404] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:30.959 [2024-09-29 16:45:52.623435] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:30.959 [2024-09-29 16:45:52.623448] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:30.959 [2024-09-29 16:45:52.626563] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:30.959 [2024-09-29 16:45:52.626613] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:30.959 BaseBdev2 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:30.959 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.960 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.219 [2024-09-29 16:45:52.635515] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:31.219 [2024-09-29 16:45:52.637455] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:31.219 [2024-09-29 16:45:52.637655] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:31.219 [2024-09-29 16:45:52.637669] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:31.219 [2024-09-29 16:45:52.637977] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:31.219 [2024-09-29 16:45:52.638109] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:31.219 [2024-09-29 16:45:52.638129] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:31.219 [2024-09-29 16:45:52.638255] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:31.219 "name": "raid_bdev1", 00:07:31.219 "uuid": "fcb5cff7-e278-446f-a4d5-131805027bb3", 00:07:31.219 "strip_size_kb": 64, 00:07:31.219 "state": "online", 00:07:31.219 "raid_level": "concat", 00:07:31.219 "superblock": true, 00:07:31.219 "num_base_bdevs": 2, 00:07:31.219 "num_base_bdevs_discovered": 2, 00:07:31.219 "num_base_bdevs_operational": 2, 00:07:31.219 "base_bdevs_list": [ 00:07:31.219 { 00:07:31.219 "name": "BaseBdev1", 00:07:31.219 "uuid": "5c428752-20ac-5953-92aa-38211dcc4a8f", 00:07:31.219 "is_configured": true, 00:07:31.219 "data_offset": 2048, 00:07:31.219 "data_size": 63488 00:07:31.219 }, 00:07:31.219 { 00:07:31.219 "name": "BaseBdev2", 00:07:31.219 "uuid": "a353b6c4-60d2-5a09-bb36-046949744975", 00:07:31.219 "is_configured": true, 00:07:31.219 "data_offset": 2048, 00:07:31.219 "data_size": 63488 00:07:31.219 } 00:07:31.219 ] 00:07:31.219 }' 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:31.219 16:45:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.478 16:45:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:31.478 16:45:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:31.478 [2024-09-29 16:45:53.127051] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.417 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.677 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:32.677 "name": "raid_bdev1", 00:07:32.677 "uuid": "fcb5cff7-e278-446f-a4d5-131805027bb3", 00:07:32.677 "strip_size_kb": 64, 00:07:32.677 "state": "online", 00:07:32.677 "raid_level": "concat", 00:07:32.677 "superblock": true, 00:07:32.678 "num_base_bdevs": 2, 00:07:32.678 "num_base_bdevs_discovered": 2, 00:07:32.678 "num_base_bdevs_operational": 2, 00:07:32.678 "base_bdevs_list": [ 00:07:32.678 { 00:07:32.678 "name": "BaseBdev1", 00:07:32.678 "uuid": "5c428752-20ac-5953-92aa-38211dcc4a8f", 00:07:32.678 "is_configured": true, 00:07:32.678 "data_offset": 2048, 00:07:32.678 "data_size": 63488 00:07:32.678 }, 00:07:32.678 { 00:07:32.678 "name": "BaseBdev2", 00:07:32.678 "uuid": "a353b6c4-60d2-5a09-bb36-046949744975", 00:07:32.678 "is_configured": true, 00:07:32.678 "data_offset": 2048, 00:07:32.678 "data_size": 63488 00:07:32.678 } 00:07:32.678 ] 00:07:32.678 }' 00:07:32.678 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:32.678 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.938 [2024-09-29 16:45:54.478779] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:32.938 [2024-09-29 16:45:54.478810] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:32.938 [2024-09-29 16:45:54.481256] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:32.938 [2024-09-29 16:45:54.481306] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:32.938 [2024-09-29 16:45:54.481341] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:32.938 [2024-09-29 16:45:54.481361] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:32.938 { 00:07:32.938 "results": [ 00:07:32.938 { 00:07:32.938 "job": "raid_bdev1", 00:07:32.938 "core_mask": "0x1", 00:07:32.938 "workload": "randrw", 00:07:32.938 "percentage": 50, 00:07:32.938 "status": "finished", 00:07:32.938 "queue_depth": 1, 00:07:32.938 "io_size": 131072, 00:07:32.938 "runtime": 1.35249, 00:07:32.938 "iops": 17833.033885647954, 00:07:32.938 "mibps": 2229.1292357059942, 00:07:32.938 "io_failed": 1, 00:07:32.938 "io_timeout": 0, 00:07:32.938 "avg_latency_us": 77.57005134444228, 00:07:32.938 "min_latency_us": 24.482096069868994, 00:07:32.938 "max_latency_us": 1380.8349344978167 00:07:32.938 } 00:07:32.938 ], 00:07:32.938 "core_count": 1 00:07:32.938 } 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73553 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 73553 ']' 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 73553 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73553 00:07:32.938 killing process with pid 73553 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73553' 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 73553 00:07:32.938 [2024-09-29 16:45:54.515548] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:32.938 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 73553 00:07:32.938 [2024-09-29 16:45:54.531724] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.bwaLzuO7Ki 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:33.199 ************************************ 00:07:33.199 00:07:33.199 real 0m3.154s 00:07:33.199 user 0m3.948s 00:07:33.199 sys 0m0.509s 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:33.199 16:45:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.199 END TEST raid_write_error_test 00:07:33.199 ************************************ 00:07:33.199 16:45:54 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:33.199 16:45:54 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:07:33.199 16:45:54 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:33.199 16:45:54 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:33.199 16:45:54 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:33.200 ************************************ 00:07:33.200 START TEST raid_state_function_test 00:07:33.200 ************************************ 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 false 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73681 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:33.200 Process raid pid: 73681 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73681' 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73681 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 73681 ']' 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:33.200 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:33.200 16:45:54 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:33.460 [2024-09-29 16:45:54.932874] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:33.460 [2024-09-29 16:45:54.933065] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:33.460 [2024-09-29 16:45:55.072820] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:33.460 [2024-09-29 16:45:55.118427] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:33.720 [2024-09-29 16:45:55.161843] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:33.720 [2024-09-29 16:45:55.161876] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.288 [2024-09-29 16:45:55.751581] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:34.288 [2024-09-29 16:45:55.751631] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:34.288 [2024-09-29 16:45:55.751649] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:34.288 [2024-09-29 16:45:55.751659] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.288 "name": "Existed_Raid", 00:07:34.288 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.288 "strip_size_kb": 0, 00:07:34.288 "state": "configuring", 00:07:34.288 "raid_level": "raid1", 00:07:34.288 "superblock": false, 00:07:34.288 "num_base_bdevs": 2, 00:07:34.288 "num_base_bdevs_discovered": 0, 00:07:34.288 "num_base_bdevs_operational": 2, 00:07:34.288 "base_bdevs_list": [ 00:07:34.288 { 00:07:34.288 "name": "BaseBdev1", 00:07:34.288 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.288 "is_configured": false, 00:07:34.288 "data_offset": 0, 00:07:34.288 "data_size": 0 00:07:34.288 }, 00:07:34.288 { 00:07:34.288 "name": "BaseBdev2", 00:07:34.288 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.288 "is_configured": false, 00:07:34.288 "data_offset": 0, 00:07:34.288 "data_size": 0 00:07:34.288 } 00:07:34.288 ] 00:07:34.288 }' 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.288 16:45:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.547 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:34.547 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.547 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.806 [2024-09-29 16:45:56.222681] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:34.806 [2024-09-29 16:45:56.222739] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.806 [2024-09-29 16:45:56.230668] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:34.806 [2024-09-29 16:45:56.230756] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:34.806 [2024-09-29 16:45:56.230811] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:34.806 [2024-09-29 16:45:56.230850] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.806 [2024-09-29 16:45:56.251410] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:34.806 BaseBdev1 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.806 [ 00:07:34.806 { 00:07:34.806 "name": "BaseBdev1", 00:07:34.806 "aliases": [ 00:07:34.806 "7563ccb1-bab4-4ad9-b4a0-a0244374548c" 00:07:34.806 ], 00:07:34.806 "product_name": "Malloc disk", 00:07:34.806 "block_size": 512, 00:07:34.806 "num_blocks": 65536, 00:07:34.806 "uuid": "7563ccb1-bab4-4ad9-b4a0-a0244374548c", 00:07:34.806 "assigned_rate_limits": { 00:07:34.806 "rw_ios_per_sec": 0, 00:07:34.806 "rw_mbytes_per_sec": 0, 00:07:34.806 "r_mbytes_per_sec": 0, 00:07:34.806 "w_mbytes_per_sec": 0 00:07:34.806 }, 00:07:34.806 "claimed": true, 00:07:34.806 "claim_type": "exclusive_write", 00:07:34.806 "zoned": false, 00:07:34.806 "supported_io_types": { 00:07:34.806 "read": true, 00:07:34.806 "write": true, 00:07:34.806 "unmap": true, 00:07:34.806 "flush": true, 00:07:34.806 "reset": true, 00:07:34.806 "nvme_admin": false, 00:07:34.806 "nvme_io": false, 00:07:34.806 "nvme_io_md": false, 00:07:34.806 "write_zeroes": true, 00:07:34.806 "zcopy": true, 00:07:34.806 "get_zone_info": false, 00:07:34.806 "zone_management": false, 00:07:34.806 "zone_append": false, 00:07:34.806 "compare": false, 00:07:34.806 "compare_and_write": false, 00:07:34.806 "abort": true, 00:07:34.806 "seek_hole": false, 00:07:34.806 "seek_data": false, 00:07:34.806 "copy": true, 00:07:34.806 "nvme_iov_md": false 00:07:34.806 }, 00:07:34.806 "memory_domains": [ 00:07:34.806 { 00:07:34.806 "dma_device_id": "system", 00:07:34.806 "dma_device_type": 1 00:07:34.806 }, 00:07:34.806 { 00:07:34.806 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.806 "dma_device_type": 2 00:07:34.806 } 00:07:34.806 ], 00:07:34.806 "driver_specific": {} 00:07:34.806 } 00:07:34.806 ] 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:34.806 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.807 "name": "Existed_Raid", 00:07:34.807 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.807 "strip_size_kb": 0, 00:07:34.807 "state": "configuring", 00:07:34.807 "raid_level": "raid1", 00:07:34.807 "superblock": false, 00:07:34.807 "num_base_bdevs": 2, 00:07:34.807 "num_base_bdevs_discovered": 1, 00:07:34.807 "num_base_bdevs_operational": 2, 00:07:34.807 "base_bdevs_list": [ 00:07:34.807 { 00:07:34.807 "name": "BaseBdev1", 00:07:34.807 "uuid": "7563ccb1-bab4-4ad9-b4a0-a0244374548c", 00:07:34.807 "is_configured": true, 00:07:34.807 "data_offset": 0, 00:07:34.807 "data_size": 65536 00:07:34.807 }, 00:07:34.807 { 00:07:34.807 "name": "BaseBdev2", 00:07:34.807 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.807 "is_configured": false, 00:07:34.807 "data_offset": 0, 00:07:34.807 "data_size": 0 00:07:34.807 } 00:07:34.807 ] 00:07:34.807 }' 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.807 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.066 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:35.066 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.066 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.066 [2024-09-29 16:45:56.738597] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:35.066 [2024-09-29 16:45:56.738683] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.325 [2024-09-29 16:45:56.750618] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:35.325 [2024-09-29 16:45:56.752557] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:35.325 [2024-09-29 16:45:56.752632] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.325 "name": "Existed_Raid", 00:07:35.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.325 "strip_size_kb": 0, 00:07:35.325 "state": "configuring", 00:07:35.325 "raid_level": "raid1", 00:07:35.325 "superblock": false, 00:07:35.325 "num_base_bdevs": 2, 00:07:35.325 "num_base_bdevs_discovered": 1, 00:07:35.325 "num_base_bdevs_operational": 2, 00:07:35.325 "base_bdevs_list": [ 00:07:35.325 { 00:07:35.325 "name": "BaseBdev1", 00:07:35.325 "uuid": "7563ccb1-bab4-4ad9-b4a0-a0244374548c", 00:07:35.325 "is_configured": true, 00:07:35.325 "data_offset": 0, 00:07:35.325 "data_size": 65536 00:07:35.325 }, 00:07:35.325 { 00:07:35.325 "name": "BaseBdev2", 00:07:35.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.325 "is_configured": false, 00:07:35.325 "data_offset": 0, 00:07:35.325 "data_size": 0 00:07:35.325 } 00:07:35.325 ] 00:07:35.325 }' 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.325 16:45:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.584 [2024-09-29 16:45:57.248531] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:35.584 [2024-09-29 16:45:57.248576] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:35.584 [2024-09-29 16:45:57.248585] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:35.584 [2024-09-29 16:45:57.248849] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:35.584 [2024-09-29 16:45:57.249030] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:35.584 [2024-09-29 16:45:57.249045] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:35.584 [2024-09-29 16:45:57.249273] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:35.584 BaseBdev2 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.584 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.843 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.843 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:35.843 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.843 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.843 [ 00:07:35.843 { 00:07:35.843 "name": "BaseBdev2", 00:07:35.843 "aliases": [ 00:07:35.843 "cea95867-85c1-49cf-ba86-74ebf2e8affb" 00:07:35.843 ], 00:07:35.843 "product_name": "Malloc disk", 00:07:35.843 "block_size": 512, 00:07:35.843 "num_blocks": 65536, 00:07:35.843 "uuid": "cea95867-85c1-49cf-ba86-74ebf2e8affb", 00:07:35.843 "assigned_rate_limits": { 00:07:35.843 "rw_ios_per_sec": 0, 00:07:35.843 "rw_mbytes_per_sec": 0, 00:07:35.843 "r_mbytes_per_sec": 0, 00:07:35.843 "w_mbytes_per_sec": 0 00:07:35.843 }, 00:07:35.843 "claimed": true, 00:07:35.843 "claim_type": "exclusive_write", 00:07:35.843 "zoned": false, 00:07:35.843 "supported_io_types": { 00:07:35.843 "read": true, 00:07:35.843 "write": true, 00:07:35.843 "unmap": true, 00:07:35.843 "flush": true, 00:07:35.843 "reset": true, 00:07:35.843 "nvme_admin": false, 00:07:35.843 "nvme_io": false, 00:07:35.843 "nvme_io_md": false, 00:07:35.843 "write_zeroes": true, 00:07:35.843 "zcopy": true, 00:07:35.843 "get_zone_info": false, 00:07:35.843 "zone_management": false, 00:07:35.843 "zone_append": false, 00:07:35.844 "compare": false, 00:07:35.844 "compare_and_write": false, 00:07:35.844 "abort": true, 00:07:35.844 "seek_hole": false, 00:07:35.844 "seek_data": false, 00:07:35.844 "copy": true, 00:07:35.844 "nvme_iov_md": false 00:07:35.844 }, 00:07:35.844 "memory_domains": [ 00:07:35.844 { 00:07:35.844 "dma_device_id": "system", 00:07:35.844 "dma_device_type": 1 00:07:35.844 }, 00:07:35.844 { 00:07:35.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.844 "dma_device_type": 2 00:07:35.844 } 00:07:35.844 ], 00:07:35.844 "driver_specific": {} 00:07:35.844 } 00:07:35.844 ] 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.844 "name": "Existed_Raid", 00:07:35.844 "uuid": "7b878962-2559-4da8-a8fe-237edd35c7c2", 00:07:35.844 "strip_size_kb": 0, 00:07:35.844 "state": "online", 00:07:35.844 "raid_level": "raid1", 00:07:35.844 "superblock": false, 00:07:35.844 "num_base_bdevs": 2, 00:07:35.844 "num_base_bdevs_discovered": 2, 00:07:35.844 "num_base_bdevs_operational": 2, 00:07:35.844 "base_bdevs_list": [ 00:07:35.844 { 00:07:35.844 "name": "BaseBdev1", 00:07:35.844 "uuid": "7563ccb1-bab4-4ad9-b4a0-a0244374548c", 00:07:35.844 "is_configured": true, 00:07:35.844 "data_offset": 0, 00:07:35.844 "data_size": 65536 00:07:35.844 }, 00:07:35.844 { 00:07:35.844 "name": "BaseBdev2", 00:07:35.844 "uuid": "cea95867-85c1-49cf-ba86-74ebf2e8affb", 00:07:35.844 "is_configured": true, 00:07:35.844 "data_offset": 0, 00:07:35.844 "data_size": 65536 00:07:35.844 } 00:07:35.844 ] 00:07:35.844 }' 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.844 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.103 [2024-09-29 16:45:57.740001] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:36.103 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.362 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:36.362 "name": "Existed_Raid", 00:07:36.362 "aliases": [ 00:07:36.362 "7b878962-2559-4da8-a8fe-237edd35c7c2" 00:07:36.362 ], 00:07:36.362 "product_name": "Raid Volume", 00:07:36.362 "block_size": 512, 00:07:36.362 "num_blocks": 65536, 00:07:36.362 "uuid": "7b878962-2559-4da8-a8fe-237edd35c7c2", 00:07:36.362 "assigned_rate_limits": { 00:07:36.362 "rw_ios_per_sec": 0, 00:07:36.362 "rw_mbytes_per_sec": 0, 00:07:36.362 "r_mbytes_per_sec": 0, 00:07:36.363 "w_mbytes_per_sec": 0 00:07:36.363 }, 00:07:36.363 "claimed": false, 00:07:36.363 "zoned": false, 00:07:36.363 "supported_io_types": { 00:07:36.363 "read": true, 00:07:36.363 "write": true, 00:07:36.363 "unmap": false, 00:07:36.363 "flush": false, 00:07:36.363 "reset": true, 00:07:36.363 "nvme_admin": false, 00:07:36.363 "nvme_io": false, 00:07:36.363 "nvme_io_md": false, 00:07:36.363 "write_zeroes": true, 00:07:36.363 "zcopy": false, 00:07:36.363 "get_zone_info": false, 00:07:36.363 "zone_management": false, 00:07:36.363 "zone_append": false, 00:07:36.363 "compare": false, 00:07:36.363 "compare_and_write": false, 00:07:36.363 "abort": false, 00:07:36.363 "seek_hole": false, 00:07:36.363 "seek_data": false, 00:07:36.363 "copy": false, 00:07:36.363 "nvme_iov_md": false 00:07:36.363 }, 00:07:36.363 "memory_domains": [ 00:07:36.363 { 00:07:36.363 "dma_device_id": "system", 00:07:36.363 "dma_device_type": 1 00:07:36.363 }, 00:07:36.363 { 00:07:36.363 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.363 "dma_device_type": 2 00:07:36.363 }, 00:07:36.363 { 00:07:36.363 "dma_device_id": "system", 00:07:36.363 "dma_device_type": 1 00:07:36.363 }, 00:07:36.363 { 00:07:36.363 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:36.363 "dma_device_type": 2 00:07:36.363 } 00:07:36.363 ], 00:07:36.363 "driver_specific": { 00:07:36.363 "raid": { 00:07:36.363 "uuid": "7b878962-2559-4da8-a8fe-237edd35c7c2", 00:07:36.363 "strip_size_kb": 0, 00:07:36.363 "state": "online", 00:07:36.363 "raid_level": "raid1", 00:07:36.363 "superblock": false, 00:07:36.363 "num_base_bdevs": 2, 00:07:36.363 "num_base_bdevs_discovered": 2, 00:07:36.363 "num_base_bdevs_operational": 2, 00:07:36.363 "base_bdevs_list": [ 00:07:36.363 { 00:07:36.363 "name": "BaseBdev1", 00:07:36.363 "uuid": "7563ccb1-bab4-4ad9-b4a0-a0244374548c", 00:07:36.363 "is_configured": true, 00:07:36.363 "data_offset": 0, 00:07:36.363 "data_size": 65536 00:07:36.363 }, 00:07:36.363 { 00:07:36.363 "name": "BaseBdev2", 00:07:36.363 "uuid": "cea95867-85c1-49cf-ba86-74ebf2e8affb", 00:07:36.363 "is_configured": true, 00:07:36.363 "data_offset": 0, 00:07:36.363 "data_size": 65536 00:07:36.363 } 00:07:36.363 ] 00:07:36.363 } 00:07:36.363 } 00:07:36.363 }' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:36.363 BaseBdev2' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.363 [2024-09-29 16:45:57.947421] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.363 16:45:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.363 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:36.363 "name": "Existed_Raid", 00:07:36.363 "uuid": "7b878962-2559-4da8-a8fe-237edd35c7c2", 00:07:36.363 "strip_size_kb": 0, 00:07:36.363 "state": "online", 00:07:36.363 "raid_level": "raid1", 00:07:36.363 "superblock": false, 00:07:36.363 "num_base_bdevs": 2, 00:07:36.363 "num_base_bdevs_discovered": 1, 00:07:36.363 "num_base_bdevs_operational": 1, 00:07:36.363 "base_bdevs_list": [ 00:07:36.363 { 00:07:36.363 "name": null, 00:07:36.363 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:36.363 "is_configured": false, 00:07:36.363 "data_offset": 0, 00:07:36.363 "data_size": 65536 00:07:36.363 }, 00:07:36.363 { 00:07:36.363 "name": "BaseBdev2", 00:07:36.363 "uuid": "cea95867-85c1-49cf-ba86-74ebf2e8affb", 00:07:36.363 "is_configured": true, 00:07:36.363 "data_offset": 0, 00:07:36.363 "data_size": 65536 00:07:36.363 } 00:07:36.363 ] 00:07:36.363 }' 00:07:36.363 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:36.363 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.932 [2024-09-29 16:45:58.405870] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:36.932 [2024-09-29 16:45:58.406007] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:36.932 [2024-09-29 16:45:58.417603] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:36.932 [2024-09-29 16:45:58.417730] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:36.932 [2024-09-29 16:45:58.417790] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.932 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73681 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 73681 ']' 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 73681 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73681 00:07:36.933 killing process with pid 73681 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73681' 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 73681 00:07:36.933 [2024-09-29 16:45:58.511187] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:36.933 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 73681 00:07:36.933 [2024-09-29 16:45:58.512168] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:37.192 00:07:37.192 real 0m3.910s 00:07:37.192 user 0m6.169s 00:07:37.192 sys 0m0.758s 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.192 ************************************ 00:07:37.192 END TEST raid_state_function_test 00:07:37.192 ************************************ 00:07:37.192 16:45:58 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:07:37.192 16:45:58 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:37.192 16:45:58 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:37.192 16:45:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:37.192 ************************************ 00:07:37.192 START TEST raid_state_function_test_sb 00:07:37.192 ************************************ 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:37.192 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73922 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73922' 00:07:37.193 Process raid pid: 73922 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73922 00:07:37.193 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 73922 ']' 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:37.193 16:45:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:37.451 [2024-09-29 16:45:58.911349] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:37.451 [2024-09-29 16:45:58.911553] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:37.451 [2024-09-29 16:45:59.055962] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:37.451 [2024-09-29 16:45:59.101367] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:37.709 [2024-09-29 16:45:59.143315] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:37.709 [2024-09-29 16:45:59.143354] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:38.276 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:38.276 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:38.276 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:38.276 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.277 [2024-09-29 16:45:59.744310] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:38.277 [2024-09-29 16:45:59.744363] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:38.277 [2024-09-29 16:45:59.744376] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:38.277 [2024-09-29 16:45:59.744387] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.277 "name": "Existed_Raid", 00:07:38.277 "uuid": "3825ea00-ee52-43c8-b756-bcd2dd382c05", 00:07:38.277 "strip_size_kb": 0, 00:07:38.277 "state": "configuring", 00:07:38.277 "raid_level": "raid1", 00:07:38.277 "superblock": true, 00:07:38.277 "num_base_bdevs": 2, 00:07:38.277 "num_base_bdevs_discovered": 0, 00:07:38.277 "num_base_bdevs_operational": 2, 00:07:38.277 "base_bdevs_list": [ 00:07:38.277 { 00:07:38.277 "name": "BaseBdev1", 00:07:38.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:38.277 "is_configured": false, 00:07:38.277 "data_offset": 0, 00:07:38.277 "data_size": 0 00:07:38.277 }, 00:07:38.277 { 00:07:38.277 "name": "BaseBdev2", 00:07:38.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:38.277 "is_configured": false, 00:07:38.277 "data_offset": 0, 00:07:38.277 "data_size": 0 00:07:38.277 } 00:07:38.277 ] 00:07:38.277 }' 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.277 16:45:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.536 [2024-09-29 16:46:00.167475] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:38.536 [2024-09-29 16:46:00.167557] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.536 [2024-09-29 16:46:00.179471] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:38.536 [2024-09-29 16:46:00.179563] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:38.536 [2024-09-29 16:46:00.179602] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:38.536 [2024-09-29 16:46:00.179626] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.536 [2024-09-29 16:46:00.200235] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:38.536 BaseBdev1 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.536 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.797 [ 00:07:38.797 { 00:07:38.797 "name": "BaseBdev1", 00:07:38.797 "aliases": [ 00:07:38.797 "7d107d0b-eca9-46dd-bc3a-be72adfd5a7e" 00:07:38.797 ], 00:07:38.797 "product_name": "Malloc disk", 00:07:38.797 "block_size": 512, 00:07:38.797 "num_blocks": 65536, 00:07:38.797 "uuid": "7d107d0b-eca9-46dd-bc3a-be72adfd5a7e", 00:07:38.797 "assigned_rate_limits": { 00:07:38.797 "rw_ios_per_sec": 0, 00:07:38.797 "rw_mbytes_per_sec": 0, 00:07:38.797 "r_mbytes_per_sec": 0, 00:07:38.797 "w_mbytes_per_sec": 0 00:07:38.797 }, 00:07:38.797 "claimed": true, 00:07:38.797 "claim_type": "exclusive_write", 00:07:38.797 "zoned": false, 00:07:38.797 "supported_io_types": { 00:07:38.797 "read": true, 00:07:38.797 "write": true, 00:07:38.797 "unmap": true, 00:07:38.797 "flush": true, 00:07:38.797 "reset": true, 00:07:38.797 "nvme_admin": false, 00:07:38.797 "nvme_io": false, 00:07:38.797 "nvme_io_md": false, 00:07:38.797 "write_zeroes": true, 00:07:38.797 "zcopy": true, 00:07:38.797 "get_zone_info": false, 00:07:38.797 "zone_management": false, 00:07:38.797 "zone_append": false, 00:07:38.797 "compare": false, 00:07:38.797 "compare_and_write": false, 00:07:38.797 "abort": true, 00:07:38.797 "seek_hole": false, 00:07:38.797 "seek_data": false, 00:07:38.797 "copy": true, 00:07:38.797 "nvme_iov_md": false 00:07:38.797 }, 00:07:38.797 "memory_domains": [ 00:07:38.797 { 00:07:38.797 "dma_device_id": "system", 00:07:38.797 "dma_device_type": 1 00:07:38.797 }, 00:07:38.797 { 00:07:38.797 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:38.797 "dma_device_type": 2 00:07:38.797 } 00:07:38.797 ], 00:07:38.797 "driver_specific": {} 00:07:38.797 } 00:07:38.797 ] 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.797 "name": "Existed_Raid", 00:07:38.797 "uuid": "f6bc5832-e148-4fa2-a6b1-4f9c54f239bc", 00:07:38.797 "strip_size_kb": 0, 00:07:38.797 "state": "configuring", 00:07:38.797 "raid_level": "raid1", 00:07:38.797 "superblock": true, 00:07:38.797 "num_base_bdevs": 2, 00:07:38.797 "num_base_bdevs_discovered": 1, 00:07:38.797 "num_base_bdevs_operational": 2, 00:07:38.797 "base_bdevs_list": [ 00:07:38.797 { 00:07:38.797 "name": "BaseBdev1", 00:07:38.797 "uuid": "7d107d0b-eca9-46dd-bc3a-be72adfd5a7e", 00:07:38.797 "is_configured": true, 00:07:38.797 "data_offset": 2048, 00:07:38.797 "data_size": 63488 00:07:38.797 }, 00:07:38.797 { 00:07:38.797 "name": "BaseBdev2", 00:07:38.797 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:38.797 "is_configured": false, 00:07:38.797 "data_offset": 0, 00:07:38.797 "data_size": 0 00:07:38.797 } 00:07:38.797 ] 00:07:38.797 }' 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.797 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.057 [2024-09-29 16:46:00.655486] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:39.057 [2024-09-29 16:46:00.655530] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.057 [2024-09-29 16:46:00.667538] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:39.057 [2024-09-29 16:46:00.669417] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:39.057 [2024-09-29 16:46:00.669497] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.057 "name": "Existed_Raid", 00:07:39.057 "uuid": "5ff167a4-fdbc-44b1-9f26-1c327f8e8ef7", 00:07:39.057 "strip_size_kb": 0, 00:07:39.057 "state": "configuring", 00:07:39.057 "raid_level": "raid1", 00:07:39.057 "superblock": true, 00:07:39.057 "num_base_bdevs": 2, 00:07:39.057 "num_base_bdevs_discovered": 1, 00:07:39.057 "num_base_bdevs_operational": 2, 00:07:39.057 "base_bdevs_list": [ 00:07:39.057 { 00:07:39.057 "name": "BaseBdev1", 00:07:39.057 "uuid": "7d107d0b-eca9-46dd-bc3a-be72adfd5a7e", 00:07:39.057 "is_configured": true, 00:07:39.057 "data_offset": 2048, 00:07:39.057 "data_size": 63488 00:07:39.057 }, 00:07:39.057 { 00:07:39.057 "name": "BaseBdev2", 00:07:39.057 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:39.057 "is_configured": false, 00:07:39.057 "data_offset": 0, 00:07:39.057 "data_size": 0 00:07:39.057 } 00:07:39.057 ] 00:07:39.057 }' 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.057 16:46:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.627 [2024-09-29 16:46:01.117107] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:39.627 [2024-09-29 16:46:01.117953] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:39.627 [2024-09-29 16:46:01.118131] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:39.627 BaseBdev2 00:07:39.627 [2024-09-29 16:46:01.119329] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.627 [2024-09-29 16:46:01.120025] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:39.627 [2024-09-29 16:46:01.120125] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:39.627 [2024-09-29 16:46:01.120530] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.627 [ 00:07:39.627 { 00:07:39.627 "name": "BaseBdev2", 00:07:39.627 "aliases": [ 00:07:39.627 "4a361377-f0e4-4696-b008-d8e21d4be1e0" 00:07:39.627 ], 00:07:39.627 "product_name": "Malloc disk", 00:07:39.627 "block_size": 512, 00:07:39.627 "num_blocks": 65536, 00:07:39.627 "uuid": "4a361377-f0e4-4696-b008-d8e21d4be1e0", 00:07:39.627 "assigned_rate_limits": { 00:07:39.627 "rw_ios_per_sec": 0, 00:07:39.627 "rw_mbytes_per_sec": 0, 00:07:39.627 "r_mbytes_per_sec": 0, 00:07:39.627 "w_mbytes_per_sec": 0 00:07:39.627 }, 00:07:39.627 "claimed": true, 00:07:39.627 "claim_type": "exclusive_write", 00:07:39.627 "zoned": false, 00:07:39.627 "supported_io_types": { 00:07:39.627 "read": true, 00:07:39.627 "write": true, 00:07:39.627 "unmap": true, 00:07:39.627 "flush": true, 00:07:39.627 "reset": true, 00:07:39.627 "nvme_admin": false, 00:07:39.627 "nvme_io": false, 00:07:39.627 "nvme_io_md": false, 00:07:39.627 "write_zeroes": true, 00:07:39.627 "zcopy": true, 00:07:39.627 "get_zone_info": false, 00:07:39.627 "zone_management": false, 00:07:39.627 "zone_append": false, 00:07:39.627 "compare": false, 00:07:39.627 "compare_and_write": false, 00:07:39.627 "abort": true, 00:07:39.627 "seek_hole": false, 00:07:39.627 "seek_data": false, 00:07:39.627 "copy": true, 00:07:39.627 "nvme_iov_md": false 00:07:39.627 }, 00:07:39.627 "memory_domains": [ 00:07:39.627 { 00:07:39.627 "dma_device_id": "system", 00:07:39.627 "dma_device_type": 1 00:07:39.627 }, 00:07:39.627 { 00:07:39.627 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.627 "dma_device_type": 2 00:07:39.627 } 00:07:39.627 ], 00:07:39.627 "driver_specific": {} 00:07:39.627 } 00:07:39.627 ] 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:39.627 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:39.628 "name": "Existed_Raid", 00:07:39.628 "uuid": "5ff167a4-fdbc-44b1-9f26-1c327f8e8ef7", 00:07:39.628 "strip_size_kb": 0, 00:07:39.628 "state": "online", 00:07:39.628 "raid_level": "raid1", 00:07:39.628 "superblock": true, 00:07:39.628 "num_base_bdevs": 2, 00:07:39.628 "num_base_bdevs_discovered": 2, 00:07:39.628 "num_base_bdevs_operational": 2, 00:07:39.628 "base_bdevs_list": [ 00:07:39.628 { 00:07:39.628 "name": "BaseBdev1", 00:07:39.628 "uuid": "7d107d0b-eca9-46dd-bc3a-be72adfd5a7e", 00:07:39.628 "is_configured": true, 00:07:39.628 "data_offset": 2048, 00:07:39.628 "data_size": 63488 00:07:39.628 }, 00:07:39.628 { 00:07:39.628 "name": "BaseBdev2", 00:07:39.628 "uuid": "4a361377-f0e4-4696-b008-d8e21d4be1e0", 00:07:39.628 "is_configured": true, 00:07:39.628 "data_offset": 2048, 00:07:39.628 "data_size": 63488 00:07:39.628 } 00:07:39.628 ] 00:07:39.628 }' 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:39.628 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.197 [2024-09-29 16:46:01.596506] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:40.197 "name": "Existed_Raid", 00:07:40.197 "aliases": [ 00:07:40.197 "5ff167a4-fdbc-44b1-9f26-1c327f8e8ef7" 00:07:40.197 ], 00:07:40.197 "product_name": "Raid Volume", 00:07:40.197 "block_size": 512, 00:07:40.197 "num_blocks": 63488, 00:07:40.197 "uuid": "5ff167a4-fdbc-44b1-9f26-1c327f8e8ef7", 00:07:40.197 "assigned_rate_limits": { 00:07:40.197 "rw_ios_per_sec": 0, 00:07:40.197 "rw_mbytes_per_sec": 0, 00:07:40.197 "r_mbytes_per_sec": 0, 00:07:40.197 "w_mbytes_per_sec": 0 00:07:40.197 }, 00:07:40.197 "claimed": false, 00:07:40.197 "zoned": false, 00:07:40.197 "supported_io_types": { 00:07:40.197 "read": true, 00:07:40.197 "write": true, 00:07:40.197 "unmap": false, 00:07:40.197 "flush": false, 00:07:40.197 "reset": true, 00:07:40.197 "nvme_admin": false, 00:07:40.197 "nvme_io": false, 00:07:40.197 "nvme_io_md": false, 00:07:40.197 "write_zeroes": true, 00:07:40.197 "zcopy": false, 00:07:40.197 "get_zone_info": false, 00:07:40.197 "zone_management": false, 00:07:40.197 "zone_append": false, 00:07:40.197 "compare": false, 00:07:40.197 "compare_and_write": false, 00:07:40.197 "abort": false, 00:07:40.197 "seek_hole": false, 00:07:40.197 "seek_data": false, 00:07:40.197 "copy": false, 00:07:40.197 "nvme_iov_md": false 00:07:40.197 }, 00:07:40.197 "memory_domains": [ 00:07:40.197 { 00:07:40.197 "dma_device_id": "system", 00:07:40.197 "dma_device_type": 1 00:07:40.197 }, 00:07:40.197 { 00:07:40.197 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:40.197 "dma_device_type": 2 00:07:40.197 }, 00:07:40.197 { 00:07:40.197 "dma_device_id": "system", 00:07:40.197 "dma_device_type": 1 00:07:40.197 }, 00:07:40.197 { 00:07:40.197 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:40.197 "dma_device_type": 2 00:07:40.197 } 00:07:40.197 ], 00:07:40.197 "driver_specific": { 00:07:40.197 "raid": { 00:07:40.197 "uuid": "5ff167a4-fdbc-44b1-9f26-1c327f8e8ef7", 00:07:40.197 "strip_size_kb": 0, 00:07:40.197 "state": "online", 00:07:40.197 "raid_level": "raid1", 00:07:40.197 "superblock": true, 00:07:40.197 "num_base_bdevs": 2, 00:07:40.197 "num_base_bdevs_discovered": 2, 00:07:40.197 "num_base_bdevs_operational": 2, 00:07:40.197 "base_bdevs_list": [ 00:07:40.197 { 00:07:40.197 "name": "BaseBdev1", 00:07:40.197 "uuid": "7d107d0b-eca9-46dd-bc3a-be72adfd5a7e", 00:07:40.197 "is_configured": true, 00:07:40.197 "data_offset": 2048, 00:07:40.197 "data_size": 63488 00:07:40.197 }, 00:07:40.197 { 00:07:40.197 "name": "BaseBdev2", 00:07:40.197 "uuid": "4a361377-f0e4-4696-b008-d8e21d4be1e0", 00:07:40.197 "is_configured": true, 00:07:40.197 "data_offset": 2048, 00:07:40.197 "data_size": 63488 00:07:40.197 } 00:07:40.197 ] 00:07:40.197 } 00:07:40.197 } 00:07:40.197 }' 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:40.197 BaseBdev2' 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:40.197 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.198 [2024-09-29 16:46:01.811914] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.198 "name": "Existed_Raid", 00:07:40.198 "uuid": "5ff167a4-fdbc-44b1-9f26-1c327f8e8ef7", 00:07:40.198 "strip_size_kb": 0, 00:07:40.198 "state": "online", 00:07:40.198 "raid_level": "raid1", 00:07:40.198 "superblock": true, 00:07:40.198 "num_base_bdevs": 2, 00:07:40.198 "num_base_bdevs_discovered": 1, 00:07:40.198 "num_base_bdevs_operational": 1, 00:07:40.198 "base_bdevs_list": [ 00:07:40.198 { 00:07:40.198 "name": null, 00:07:40.198 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:40.198 "is_configured": false, 00:07:40.198 "data_offset": 0, 00:07:40.198 "data_size": 63488 00:07:40.198 }, 00:07:40.198 { 00:07:40.198 "name": "BaseBdev2", 00:07:40.198 "uuid": "4a361377-f0e4-4696-b008-d8e21d4be1e0", 00:07:40.198 "is_configured": true, 00:07:40.198 "data_offset": 2048, 00:07:40.198 "data_size": 63488 00:07:40.198 } 00:07:40.198 ] 00:07:40.198 }' 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.198 16:46:01 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.767 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.768 [2024-09-29 16:46:02.254510] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:40.768 [2024-09-29 16:46:02.254677] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:40.768 [2024-09-29 16:46:02.266216] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:40.768 [2024-09-29 16:46:02.266348] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:40.768 [2024-09-29 16:46:02.266401] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73922 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 73922 ']' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 73922 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73922 00:07:40.768 killing process with pid 73922 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73922' 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 73922 00:07:40.768 [2024-09-29 16:46:02.366603] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:40.768 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 73922 00:07:40.768 [2024-09-29 16:46:02.367589] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:41.028 16:46:02 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:41.028 00:07:41.028 real 0m3.782s 00:07:41.028 user 0m5.927s 00:07:41.028 sys 0m0.745s 00:07:41.028 ************************************ 00:07:41.028 END TEST raid_state_function_test_sb 00:07:41.028 ************************************ 00:07:41.028 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:41.028 16:46:02 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:41.028 16:46:02 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:07:41.028 16:46:02 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:41.028 16:46:02 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:41.028 16:46:02 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:41.028 ************************************ 00:07:41.028 START TEST raid_superblock_test 00:07:41.028 ************************************ 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74158 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74158 00:07:41.028 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 74158 ']' 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:41.028 16:46:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.288 [2024-09-29 16:46:02.768950] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:41.288 [2024-09-29 16:46:02.769172] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74158 ] 00:07:41.288 [2024-09-29 16:46:02.912415] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:41.288 [2024-09-29 16:46:02.958219] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:41.564 [2024-09-29 16:46:03.000989] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:41.564 [2024-09-29 16:46:03.001072] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.167 malloc1 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.167 [2024-09-29 16:46:03.595423] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:42.167 [2024-09-29 16:46:03.595488] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:42.167 [2024-09-29 16:46:03.595510] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:42.167 [2024-09-29 16:46:03.595526] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:42.167 [2024-09-29 16:46:03.597639] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:42.167 [2024-09-29 16:46:03.597678] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:42.167 pt1 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.167 malloc2 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.167 [2024-09-29 16:46:03.638124] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:42.167 [2024-09-29 16:46:03.638318] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:42.167 [2024-09-29 16:46:03.638398] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:42.167 [2024-09-29 16:46:03.638502] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:42.167 [2024-09-29 16:46:03.643381] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:42.167 [2024-09-29 16:46:03.643535] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:42.167 pt2 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.167 [2024-09-29 16:46:03.651882] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:42.167 [2024-09-29 16:46:03.654857] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:42.167 [2024-09-29 16:46:03.655129] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:42.167 [2024-09-29 16:46:03.655236] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:42.167 [2024-09-29 16:46:03.655736] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:42.167 [2024-09-29 16:46:03.656019] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:42.167 [2024-09-29 16:46:03.656088] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:42.167 [2024-09-29 16:46:03.656378] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:42.167 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:42.168 "name": "raid_bdev1", 00:07:42.168 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:42.168 "strip_size_kb": 0, 00:07:42.168 "state": "online", 00:07:42.168 "raid_level": "raid1", 00:07:42.168 "superblock": true, 00:07:42.168 "num_base_bdevs": 2, 00:07:42.168 "num_base_bdevs_discovered": 2, 00:07:42.168 "num_base_bdevs_operational": 2, 00:07:42.168 "base_bdevs_list": [ 00:07:42.168 { 00:07:42.168 "name": "pt1", 00:07:42.168 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:42.168 "is_configured": true, 00:07:42.168 "data_offset": 2048, 00:07:42.168 "data_size": 63488 00:07:42.168 }, 00:07:42.168 { 00:07:42.168 "name": "pt2", 00:07:42.168 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:42.168 "is_configured": true, 00:07:42.168 "data_offset": 2048, 00:07:42.168 "data_size": 63488 00:07:42.168 } 00:07:42.168 ] 00:07:42.168 }' 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:42.168 16:46:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.427 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.427 [2024-09-29 16:46:04.091894] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:42.686 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.686 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:42.686 "name": "raid_bdev1", 00:07:42.686 "aliases": [ 00:07:42.686 "af60d64d-4ed6-41ee-90df-9d991c0adc93" 00:07:42.686 ], 00:07:42.686 "product_name": "Raid Volume", 00:07:42.686 "block_size": 512, 00:07:42.686 "num_blocks": 63488, 00:07:42.686 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:42.686 "assigned_rate_limits": { 00:07:42.686 "rw_ios_per_sec": 0, 00:07:42.686 "rw_mbytes_per_sec": 0, 00:07:42.686 "r_mbytes_per_sec": 0, 00:07:42.687 "w_mbytes_per_sec": 0 00:07:42.687 }, 00:07:42.687 "claimed": false, 00:07:42.687 "zoned": false, 00:07:42.687 "supported_io_types": { 00:07:42.687 "read": true, 00:07:42.687 "write": true, 00:07:42.687 "unmap": false, 00:07:42.687 "flush": false, 00:07:42.687 "reset": true, 00:07:42.687 "nvme_admin": false, 00:07:42.687 "nvme_io": false, 00:07:42.687 "nvme_io_md": false, 00:07:42.687 "write_zeroes": true, 00:07:42.687 "zcopy": false, 00:07:42.687 "get_zone_info": false, 00:07:42.687 "zone_management": false, 00:07:42.687 "zone_append": false, 00:07:42.687 "compare": false, 00:07:42.687 "compare_and_write": false, 00:07:42.687 "abort": false, 00:07:42.687 "seek_hole": false, 00:07:42.687 "seek_data": false, 00:07:42.687 "copy": false, 00:07:42.687 "nvme_iov_md": false 00:07:42.687 }, 00:07:42.687 "memory_domains": [ 00:07:42.687 { 00:07:42.687 "dma_device_id": "system", 00:07:42.687 "dma_device_type": 1 00:07:42.687 }, 00:07:42.687 { 00:07:42.687 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:42.687 "dma_device_type": 2 00:07:42.687 }, 00:07:42.687 { 00:07:42.687 "dma_device_id": "system", 00:07:42.687 "dma_device_type": 1 00:07:42.687 }, 00:07:42.687 { 00:07:42.687 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:42.687 "dma_device_type": 2 00:07:42.687 } 00:07:42.687 ], 00:07:42.687 "driver_specific": { 00:07:42.687 "raid": { 00:07:42.687 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:42.687 "strip_size_kb": 0, 00:07:42.687 "state": "online", 00:07:42.687 "raid_level": "raid1", 00:07:42.687 "superblock": true, 00:07:42.687 "num_base_bdevs": 2, 00:07:42.687 "num_base_bdevs_discovered": 2, 00:07:42.687 "num_base_bdevs_operational": 2, 00:07:42.687 "base_bdevs_list": [ 00:07:42.687 { 00:07:42.687 "name": "pt1", 00:07:42.687 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:42.687 "is_configured": true, 00:07:42.687 "data_offset": 2048, 00:07:42.687 "data_size": 63488 00:07:42.687 }, 00:07:42.687 { 00:07:42.687 "name": "pt2", 00:07:42.687 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:42.687 "is_configured": true, 00:07:42.687 "data_offset": 2048, 00:07:42.687 "data_size": 63488 00:07:42.687 } 00:07:42.687 ] 00:07:42.687 } 00:07:42.687 } 00:07:42.687 }' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:42.687 pt2' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.687 [2024-09-29 16:46:04.315435] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=af60d64d-4ed6-41ee-90df-9d991c0adc93 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z af60d64d-4ed6-41ee-90df-9d991c0adc93 ']' 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.687 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.687 [2024-09-29 16:46:04.355151] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:42.687 [2024-09-29 16:46:04.355175] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:42.687 [2024-09-29 16:46:04.355237] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:42.687 [2024-09-29 16:46:04.355302] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:42.687 [2024-09-29 16:46:04.355313] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:42.947 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.948 [2024-09-29 16:46:04.486922] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:42.948 [2024-09-29 16:46:04.488704] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:42.948 [2024-09-29 16:46:04.488771] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:42.948 [2024-09-29 16:46:04.488825] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:42.948 [2024-09-29 16:46:04.488841] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:42.948 [2024-09-29 16:46:04.488849] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:42.948 request: 00:07:42.948 { 00:07:42.948 "name": "raid_bdev1", 00:07:42.948 "raid_level": "raid1", 00:07:42.948 "base_bdevs": [ 00:07:42.948 "malloc1", 00:07:42.948 "malloc2" 00:07:42.948 ], 00:07:42.948 "superblock": false, 00:07:42.948 "method": "bdev_raid_create", 00:07:42.948 "req_id": 1 00:07:42.948 } 00:07:42.948 Got JSON-RPC error response 00:07:42.948 response: 00:07:42.948 { 00:07:42.948 "code": -17, 00:07:42.948 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:42.948 } 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.948 [2024-09-29 16:46:04.538815] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:42.948 [2024-09-29 16:46:04.538905] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:42.948 [2024-09-29 16:46:04.538952] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:42.948 [2024-09-29 16:46:04.538981] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:42.948 [2024-09-29 16:46:04.541079] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:42.948 [2024-09-29 16:46:04.541146] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:42.948 [2024-09-29 16:46:04.541250] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:42.948 [2024-09-29 16:46:04.541332] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:42.948 pt1 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:42.948 "name": "raid_bdev1", 00:07:42.948 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:42.948 "strip_size_kb": 0, 00:07:42.948 "state": "configuring", 00:07:42.948 "raid_level": "raid1", 00:07:42.948 "superblock": true, 00:07:42.948 "num_base_bdevs": 2, 00:07:42.948 "num_base_bdevs_discovered": 1, 00:07:42.948 "num_base_bdevs_operational": 2, 00:07:42.948 "base_bdevs_list": [ 00:07:42.948 { 00:07:42.948 "name": "pt1", 00:07:42.948 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:42.948 "is_configured": true, 00:07:42.948 "data_offset": 2048, 00:07:42.948 "data_size": 63488 00:07:42.948 }, 00:07:42.948 { 00:07:42.948 "name": null, 00:07:42.948 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:42.948 "is_configured": false, 00:07:42.948 "data_offset": 2048, 00:07:42.948 "data_size": 63488 00:07:42.948 } 00:07:42.948 ] 00:07:42.948 }' 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:42.948 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.516 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:43.516 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:43.516 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:43.516 16:46:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:43.516 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.516 16:46:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.516 [2024-09-29 16:46:05.006020] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:43.516 [2024-09-29 16:46:05.006073] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:43.516 [2024-09-29 16:46:05.006092] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:43.516 [2024-09-29 16:46:05.006100] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:43.516 [2024-09-29 16:46:05.006448] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:43.516 [2024-09-29 16:46:05.006464] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:43.516 [2024-09-29 16:46:05.006524] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:43.516 [2024-09-29 16:46:05.006541] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:43.516 [2024-09-29 16:46:05.006623] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:43.516 [2024-09-29 16:46:05.006630] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:43.516 [2024-09-29 16:46:05.006900] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:43.516 [2024-09-29 16:46:05.007012] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:43.516 [2024-09-29 16:46:05.007025] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:43.516 [2024-09-29 16:46:05.007122] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:43.516 pt2 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:43.516 "name": "raid_bdev1", 00:07:43.516 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:43.516 "strip_size_kb": 0, 00:07:43.516 "state": "online", 00:07:43.516 "raid_level": "raid1", 00:07:43.516 "superblock": true, 00:07:43.516 "num_base_bdevs": 2, 00:07:43.516 "num_base_bdevs_discovered": 2, 00:07:43.516 "num_base_bdevs_operational": 2, 00:07:43.516 "base_bdevs_list": [ 00:07:43.516 { 00:07:43.516 "name": "pt1", 00:07:43.516 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:43.516 "is_configured": true, 00:07:43.516 "data_offset": 2048, 00:07:43.516 "data_size": 63488 00:07:43.516 }, 00:07:43.516 { 00:07:43.516 "name": "pt2", 00:07:43.516 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:43.516 "is_configured": true, 00:07:43.516 "data_offset": 2048, 00:07:43.516 "data_size": 63488 00:07:43.516 } 00:07:43.516 ] 00:07:43.516 }' 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:43.516 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.775 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:43.775 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:43.775 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:43.775 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:43.775 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:43.775 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:43.775 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:43.776 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:43.776 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:43.776 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.776 [2024-09-29 16:46:05.445516] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:44.035 "name": "raid_bdev1", 00:07:44.035 "aliases": [ 00:07:44.035 "af60d64d-4ed6-41ee-90df-9d991c0adc93" 00:07:44.035 ], 00:07:44.035 "product_name": "Raid Volume", 00:07:44.035 "block_size": 512, 00:07:44.035 "num_blocks": 63488, 00:07:44.035 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:44.035 "assigned_rate_limits": { 00:07:44.035 "rw_ios_per_sec": 0, 00:07:44.035 "rw_mbytes_per_sec": 0, 00:07:44.035 "r_mbytes_per_sec": 0, 00:07:44.035 "w_mbytes_per_sec": 0 00:07:44.035 }, 00:07:44.035 "claimed": false, 00:07:44.035 "zoned": false, 00:07:44.035 "supported_io_types": { 00:07:44.035 "read": true, 00:07:44.035 "write": true, 00:07:44.035 "unmap": false, 00:07:44.035 "flush": false, 00:07:44.035 "reset": true, 00:07:44.035 "nvme_admin": false, 00:07:44.035 "nvme_io": false, 00:07:44.035 "nvme_io_md": false, 00:07:44.035 "write_zeroes": true, 00:07:44.035 "zcopy": false, 00:07:44.035 "get_zone_info": false, 00:07:44.035 "zone_management": false, 00:07:44.035 "zone_append": false, 00:07:44.035 "compare": false, 00:07:44.035 "compare_and_write": false, 00:07:44.035 "abort": false, 00:07:44.035 "seek_hole": false, 00:07:44.035 "seek_data": false, 00:07:44.035 "copy": false, 00:07:44.035 "nvme_iov_md": false 00:07:44.035 }, 00:07:44.035 "memory_domains": [ 00:07:44.035 { 00:07:44.035 "dma_device_id": "system", 00:07:44.035 "dma_device_type": 1 00:07:44.035 }, 00:07:44.035 { 00:07:44.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.035 "dma_device_type": 2 00:07:44.035 }, 00:07:44.035 { 00:07:44.035 "dma_device_id": "system", 00:07:44.035 "dma_device_type": 1 00:07:44.035 }, 00:07:44.035 { 00:07:44.035 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.035 "dma_device_type": 2 00:07:44.035 } 00:07:44.035 ], 00:07:44.035 "driver_specific": { 00:07:44.035 "raid": { 00:07:44.035 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:44.035 "strip_size_kb": 0, 00:07:44.035 "state": "online", 00:07:44.035 "raid_level": "raid1", 00:07:44.035 "superblock": true, 00:07:44.035 "num_base_bdevs": 2, 00:07:44.035 "num_base_bdevs_discovered": 2, 00:07:44.035 "num_base_bdevs_operational": 2, 00:07:44.035 "base_bdevs_list": [ 00:07:44.035 { 00:07:44.035 "name": "pt1", 00:07:44.035 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:44.035 "is_configured": true, 00:07:44.035 "data_offset": 2048, 00:07:44.035 "data_size": 63488 00:07:44.035 }, 00:07:44.035 { 00:07:44.035 "name": "pt2", 00:07:44.035 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.035 "is_configured": true, 00:07:44.035 "data_offset": 2048, 00:07:44.035 "data_size": 63488 00:07:44.035 } 00:07:44.035 ] 00:07:44.035 } 00:07:44.035 } 00:07:44.035 }' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:44.035 pt2' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.035 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:44.036 [2024-09-29 16:46:05.649121] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' af60d64d-4ed6-41ee-90df-9d991c0adc93 '!=' af60d64d-4ed6-41ee-90df-9d991c0adc93 ']' 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.036 [2024-09-29 16:46:05.696867] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.036 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.294 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.294 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.294 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.294 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.294 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.294 "name": "raid_bdev1", 00:07:44.294 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:44.294 "strip_size_kb": 0, 00:07:44.294 "state": "online", 00:07:44.294 "raid_level": "raid1", 00:07:44.294 "superblock": true, 00:07:44.294 "num_base_bdevs": 2, 00:07:44.294 "num_base_bdevs_discovered": 1, 00:07:44.294 "num_base_bdevs_operational": 1, 00:07:44.294 "base_bdevs_list": [ 00:07:44.294 { 00:07:44.294 "name": null, 00:07:44.294 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:44.294 "is_configured": false, 00:07:44.294 "data_offset": 0, 00:07:44.294 "data_size": 63488 00:07:44.294 }, 00:07:44.294 { 00:07:44.294 "name": "pt2", 00:07:44.294 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.294 "is_configured": true, 00:07:44.294 "data_offset": 2048, 00:07:44.294 "data_size": 63488 00:07:44.294 } 00:07:44.294 ] 00:07:44.294 }' 00:07:44.294 16:46:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.294 16:46:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.553 [2024-09-29 16:46:06.152036] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:44.553 [2024-09-29 16:46:06.152106] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:44.553 [2024-09-29 16:46:06.152196] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:44.553 [2024-09-29 16:46:06.152291] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:44.553 [2024-09-29 16:46:06.152340] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.553 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.812 [2024-09-29 16:46:06.227896] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:44.812 [2024-09-29 16:46:06.227964] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.812 [2024-09-29 16:46:06.227986] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:07:44.812 [2024-09-29 16:46:06.227995] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.812 [2024-09-29 16:46:06.230169] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.812 [2024-09-29 16:46:06.230207] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:44.812 [2024-09-29 16:46:06.230279] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:44.812 [2024-09-29 16:46:06.230310] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:44.812 [2024-09-29 16:46:06.230383] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:07:44.812 [2024-09-29 16:46:06.230391] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:44.812 [2024-09-29 16:46:06.230614] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:44.812 [2024-09-29 16:46:06.230725] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:07:44.812 [2024-09-29 16:46:06.230769] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:07:44.812 [2024-09-29 16:46:06.230880] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:44.812 pt2 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.812 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.812 "name": "raid_bdev1", 00:07:44.812 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:44.812 "strip_size_kb": 0, 00:07:44.812 "state": "online", 00:07:44.812 "raid_level": "raid1", 00:07:44.812 "superblock": true, 00:07:44.812 "num_base_bdevs": 2, 00:07:44.812 "num_base_bdevs_discovered": 1, 00:07:44.812 "num_base_bdevs_operational": 1, 00:07:44.812 "base_bdevs_list": [ 00:07:44.812 { 00:07:44.812 "name": null, 00:07:44.812 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:44.813 "is_configured": false, 00:07:44.813 "data_offset": 2048, 00:07:44.813 "data_size": 63488 00:07:44.813 }, 00:07:44.813 { 00:07:44.813 "name": "pt2", 00:07:44.813 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:44.813 "is_configured": true, 00:07:44.813 "data_offset": 2048, 00:07:44.813 "data_size": 63488 00:07:44.813 } 00:07:44.813 ] 00:07:44.813 }' 00:07:44.813 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.813 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.071 [2024-09-29 16:46:06.695127] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:45.071 [2024-09-29 16:46:06.695196] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:45.071 [2024-09-29 16:46:06.695274] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:45.071 [2024-09-29 16:46:06.695330] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:45.071 [2024-09-29 16:46:06.695393] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:07:45.071 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.335 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:07:45.335 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:07:45.335 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:07:45.335 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:45.335 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.335 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.335 [2024-09-29 16:46:06.755039] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:45.335 [2024-09-29 16:46:06.755129] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:45.335 [2024-09-29 16:46:06.755186] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:07:45.335 [2024-09-29 16:46:06.755229] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:45.335 [2024-09-29 16:46:06.757351] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:45.336 [2024-09-29 16:46:06.757422] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:45.336 [2024-09-29 16:46:06.757506] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:45.336 [2024-09-29 16:46:06.757576] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:45.336 [2024-09-29 16:46:06.757755] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:07:45.336 [2024-09-29 16:46:06.757813] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:45.336 [2024-09-29 16:46:06.757852] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:07:45.336 [2024-09-29 16:46:06.757946] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:45.336 [2024-09-29 16:46:06.758050] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:07:45.336 [2024-09-29 16:46:06.758090] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:45.336 [2024-09-29 16:46:06.758337] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:07:45.336 [2024-09-29 16:46:06.758496] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:07:45.336 [2024-09-29 16:46:06.758537] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:07:45.336 [2024-09-29 16:46:06.758695] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:45.336 pt1 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:45.336 "name": "raid_bdev1", 00:07:45.336 "uuid": "af60d64d-4ed6-41ee-90df-9d991c0adc93", 00:07:45.336 "strip_size_kb": 0, 00:07:45.336 "state": "online", 00:07:45.336 "raid_level": "raid1", 00:07:45.336 "superblock": true, 00:07:45.336 "num_base_bdevs": 2, 00:07:45.336 "num_base_bdevs_discovered": 1, 00:07:45.336 "num_base_bdevs_operational": 1, 00:07:45.336 "base_bdevs_list": [ 00:07:45.336 { 00:07:45.336 "name": null, 00:07:45.336 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:45.336 "is_configured": false, 00:07:45.336 "data_offset": 2048, 00:07:45.336 "data_size": 63488 00:07:45.336 }, 00:07:45.336 { 00:07:45.336 "name": "pt2", 00:07:45.336 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:45.336 "is_configured": true, 00:07:45.336 "data_offset": 2048, 00:07:45.336 "data_size": 63488 00:07:45.336 } 00:07:45.336 ] 00:07:45.336 }' 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:45.336 16:46:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.597 [2024-09-29 16:46:07.238525] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:45.597 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.856 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' af60d64d-4ed6-41ee-90df-9d991c0adc93 '!=' af60d64d-4ed6-41ee-90df-9d991c0adc93 ']' 00:07:45.856 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74158 00:07:45.856 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 74158 ']' 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 74158 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74158 00:07:45.857 killing process with pid 74158 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74158' 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 74158 00:07:45.857 [2024-09-29 16:46:07.319539] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:45.857 [2024-09-29 16:46:07.319621] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:45.857 [2024-09-29 16:46:07.319670] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:45.857 [2024-09-29 16:46:07.319679] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:07:45.857 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 74158 00:07:45.857 [2024-09-29 16:46:07.341983] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:46.116 16:46:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:46.116 00:07:46.116 real 0m4.905s 00:07:46.116 user 0m8.031s 00:07:46.116 sys 0m0.978s 00:07:46.116 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:46.116 ************************************ 00:07:46.116 END TEST raid_superblock_test 00:07:46.116 ************************************ 00:07:46.116 16:46:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.116 16:46:07 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:07:46.116 16:46:07 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:46.116 16:46:07 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:46.116 16:46:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:46.116 ************************************ 00:07:46.116 START TEST raid_read_error_test 00:07:46.116 ************************************ 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 read 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.7NzAItGc1o 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74477 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74477 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 74477 ']' 00:07:46.116 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:46.116 16:46:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.116 [2024-09-29 16:46:07.755809] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:46.116 [2024-09-29 16:46:07.755934] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74477 ] 00:07:46.376 [2024-09-29 16:46:07.899466] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:46.376 [2024-09-29 16:46:07.944612] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:46.376 [2024-09-29 16:46:07.986404] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:46.376 [2024-09-29 16:46:07.986521] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.944 BaseBdev1_malloc 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.944 true 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.944 [2024-09-29 16:46:08.604289] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:46.944 [2024-09-29 16:46:08.604351] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:46.944 [2024-09-29 16:46:08.604386] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:46.944 [2024-09-29 16:46:08.604396] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:46.944 [2024-09-29 16:46:08.606574] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:46.944 [2024-09-29 16:46:08.606610] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:46.944 BaseBdev1 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.944 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.204 BaseBdev2_malloc 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.204 true 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.204 [2024-09-29 16:46:08.652224] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:47.204 [2024-09-29 16:46:08.652346] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:47.204 [2024-09-29 16:46:08.652369] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:47.204 [2024-09-29 16:46:08.652377] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:47.204 [2024-09-29 16:46:08.654399] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:47.204 [2024-09-29 16:46:08.654436] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:47.204 BaseBdev2 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.204 [2024-09-29 16:46:08.664279] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:47.204 [2024-09-29 16:46:08.666077] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:47.204 [2024-09-29 16:46:08.666253] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:47.204 [2024-09-29 16:46:08.666271] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:47.204 [2024-09-29 16:46:08.666518] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:47.204 [2024-09-29 16:46:08.666632] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:47.204 [2024-09-29 16:46:08.666643] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:47.204 [2024-09-29 16:46:08.666790] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:47.204 "name": "raid_bdev1", 00:07:47.204 "uuid": "89a6263b-e2a2-4d71-906c-c5c42592a685", 00:07:47.204 "strip_size_kb": 0, 00:07:47.204 "state": "online", 00:07:47.204 "raid_level": "raid1", 00:07:47.204 "superblock": true, 00:07:47.204 "num_base_bdevs": 2, 00:07:47.204 "num_base_bdevs_discovered": 2, 00:07:47.204 "num_base_bdevs_operational": 2, 00:07:47.204 "base_bdevs_list": [ 00:07:47.204 { 00:07:47.204 "name": "BaseBdev1", 00:07:47.204 "uuid": "9a9e36d1-737b-5fda-939a-254acc00999b", 00:07:47.204 "is_configured": true, 00:07:47.204 "data_offset": 2048, 00:07:47.204 "data_size": 63488 00:07:47.204 }, 00:07:47.204 { 00:07:47.204 "name": "BaseBdev2", 00:07:47.204 "uuid": "b0110c64-6a03-5b3f-b72a-1853ca31a798", 00:07:47.204 "is_configured": true, 00:07:47.204 "data_offset": 2048, 00:07:47.204 "data_size": 63488 00:07:47.204 } 00:07:47.204 ] 00:07:47.204 }' 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:47.204 16:46:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.464 16:46:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:47.464 16:46:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:47.723 [2024-09-29 16:46:09.159808] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.661 "name": "raid_bdev1", 00:07:48.661 "uuid": "89a6263b-e2a2-4d71-906c-c5c42592a685", 00:07:48.661 "strip_size_kb": 0, 00:07:48.661 "state": "online", 00:07:48.661 "raid_level": "raid1", 00:07:48.661 "superblock": true, 00:07:48.661 "num_base_bdevs": 2, 00:07:48.661 "num_base_bdevs_discovered": 2, 00:07:48.661 "num_base_bdevs_operational": 2, 00:07:48.661 "base_bdevs_list": [ 00:07:48.661 { 00:07:48.661 "name": "BaseBdev1", 00:07:48.661 "uuid": "9a9e36d1-737b-5fda-939a-254acc00999b", 00:07:48.661 "is_configured": true, 00:07:48.661 "data_offset": 2048, 00:07:48.661 "data_size": 63488 00:07:48.661 }, 00:07:48.661 { 00:07:48.661 "name": "BaseBdev2", 00:07:48.661 "uuid": "b0110c64-6a03-5b3f-b72a-1853ca31a798", 00:07:48.661 "is_configured": true, 00:07:48.661 "data_offset": 2048, 00:07:48.661 "data_size": 63488 00:07:48.661 } 00:07:48.661 ] 00:07:48.661 }' 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.661 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.920 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:48.920 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.920 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.920 [2024-09-29 16:46:10.563378] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:48.920 [2024-09-29 16:46:10.563463] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:48.920 [2024-09-29 16:46:10.565930] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:48.920 [2024-09-29 16:46:10.566022] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:48.920 [2024-09-29 16:46:10.566147] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:48.920 [2024-09-29 16:46:10.566202] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:48.920 { 00:07:48.920 "results": [ 00:07:48.920 { 00:07:48.920 "job": "raid_bdev1", 00:07:48.920 "core_mask": "0x1", 00:07:48.920 "workload": "randrw", 00:07:48.920 "percentage": 50, 00:07:48.920 "status": "finished", 00:07:48.920 "queue_depth": 1, 00:07:48.920 "io_size": 131072, 00:07:48.920 "runtime": 1.404655, 00:07:48.920 "iops": 20122.378804759886, 00:07:48.920 "mibps": 2515.297350594986, 00:07:48.920 "io_failed": 0, 00:07:48.920 "io_timeout": 0, 00:07:48.920 "avg_latency_us": 47.22588020272885, 00:07:48.920 "min_latency_us": 21.463755458515283, 00:07:48.920 "max_latency_us": 1345.0620087336245 00:07:48.920 } 00:07:48.920 ], 00:07:48.920 "core_count": 1 00:07:48.920 } 00:07:48.921 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.921 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74477 00:07:48.921 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 74477 ']' 00:07:48.921 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 74477 00:07:48.921 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:07:48.921 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:48.921 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74477 00:07:49.180 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:49.180 killing process with pid 74477 00:07:49.180 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:49.180 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74477' 00:07:49.180 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 74477 00:07:49.180 [2024-09-29 16:46:10.613694] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:49.180 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 74477 00:07:49.180 [2024-09-29 16:46:10.628972] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.7NzAItGc1o 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:49.440 00:07:49.440 real 0m3.209s 00:07:49.440 user 0m4.081s 00:07:49.440 sys 0m0.484s 00:07:49.440 ************************************ 00:07:49.440 END TEST raid_read_error_test 00:07:49.440 ************************************ 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:49.440 16:46:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.440 16:46:10 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:07:49.440 16:46:10 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:49.440 16:46:10 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:49.440 16:46:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:49.440 ************************************ 00:07:49.440 START TEST raid_write_error_test 00:07:49.440 ************************************ 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 write 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.l0Dt80eNQN 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74606 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74606 00:07:49.440 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 74606 ']' 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:49.440 16:46:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.440 [2024-09-29 16:46:11.037346] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:49.440 [2024-09-29 16:46:11.037484] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74606 ] 00:07:49.700 [2024-09-29 16:46:11.183751] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:49.700 [2024-09-29 16:46:11.229573] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:49.700 [2024-09-29 16:46:11.271551] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:49.700 [2024-09-29 16:46:11.271675] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.269 BaseBdev1_malloc 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.269 true 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.269 [2024-09-29 16:46:11.893576] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:50.269 [2024-09-29 16:46:11.893678] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:50.269 [2024-09-29 16:46:11.893726] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:50.269 [2024-09-29 16:46:11.893767] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:50.269 [2024-09-29 16:46:11.895915] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:50.269 [2024-09-29 16:46:11.896001] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:50.269 BaseBdev1 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.269 BaseBdev2_malloc 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.269 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.529 true 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.529 [2024-09-29 16:46:11.951451] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:50.529 [2024-09-29 16:46:11.951588] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:50.529 [2024-09-29 16:46:11.951651] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:50.529 [2024-09-29 16:46:11.951703] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:50.529 [2024-09-29 16:46:11.954832] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:50.529 [2024-09-29 16:46:11.954929] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:50.529 BaseBdev2 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.529 [2024-09-29 16:46:11.963632] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:50.529 [2024-09-29 16:46:11.965690] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:50.529 [2024-09-29 16:46:11.965938] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:50.529 [2024-09-29 16:46:11.966006] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:07:50.529 [2024-09-29 16:46:11.966333] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:50.529 [2024-09-29 16:46:11.966535] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:50.529 [2024-09-29 16:46:11.966555] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:50.529 [2024-09-29 16:46:11.966686] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.529 16:46:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.529 16:46:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:50.529 "name": "raid_bdev1", 00:07:50.530 "uuid": "34b732a6-100b-456e-8373-955ca3eb05f7", 00:07:50.530 "strip_size_kb": 0, 00:07:50.530 "state": "online", 00:07:50.530 "raid_level": "raid1", 00:07:50.530 "superblock": true, 00:07:50.530 "num_base_bdevs": 2, 00:07:50.530 "num_base_bdevs_discovered": 2, 00:07:50.530 "num_base_bdevs_operational": 2, 00:07:50.530 "base_bdevs_list": [ 00:07:50.530 { 00:07:50.530 "name": "BaseBdev1", 00:07:50.530 "uuid": "44846dc7-2448-530b-9ead-a418dda69af1", 00:07:50.530 "is_configured": true, 00:07:50.530 "data_offset": 2048, 00:07:50.530 "data_size": 63488 00:07:50.530 }, 00:07:50.530 { 00:07:50.530 "name": "BaseBdev2", 00:07:50.530 "uuid": "0b6bd919-2d83-58bd-9881-cad096b65aab", 00:07:50.530 "is_configured": true, 00:07:50.530 "data_offset": 2048, 00:07:50.530 "data_size": 63488 00:07:50.530 } 00:07:50.530 ] 00:07:50.530 }' 00:07:50.530 16:46:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:50.530 16:46:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.788 16:46:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:50.788 16:46:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:51.051 [2024-09-29 16:46:12.487291] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.987 [2024-09-29 16:46:13.407297] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:07:51.987 [2024-09-29 16:46:13.407430] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:51.987 [2024-09-29 16:46:13.407666] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002530 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.987 "name": "raid_bdev1", 00:07:51.987 "uuid": "34b732a6-100b-456e-8373-955ca3eb05f7", 00:07:51.987 "strip_size_kb": 0, 00:07:51.987 "state": "online", 00:07:51.987 "raid_level": "raid1", 00:07:51.987 "superblock": true, 00:07:51.987 "num_base_bdevs": 2, 00:07:51.987 "num_base_bdevs_discovered": 1, 00:07:51.987 "num_base_bdevs_operational": 1, 00:07:51.987 "base_bdevs_list": [ 00:07:51.987 { 00:07:51.987 "name": null, 00:07:51.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.987 "is_configured": false, 00:07:51.987 "data_offset": 0, 00:07:51.987 "data_size": 63488 00:07:51.987 }, 00:07:51.987 { 00:07:51.987 "name": "BaseBdev2", 00:07:51.987 "uuid": "0b6bd919-2d83-58bd-9881-cad096b65aab", 00:07:51.987 "is_configured": true, 00:07:51.987 "data_offset": 2048, 00:07:51.987 "data_size": 63488 00:07:51.987 } 00:07:51.987 ] 00:07:51.987 }' 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.987 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.258 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:52.258 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.258 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.258 [2024-09-29 16:46:13.868131] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:52.258 [2024-09-29 16:46:13.868167] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:52.258 [2024-09-29 16:46:13.870581] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:52.258 [2024-09-29 16:46:13.870629] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:52.258 [2024-09-29 16:46:13.870678] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:52.258 [2024-09-29 16:46:13.870689] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:52.258 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.258 { 00:07:52.258 "results": [ 00:07:52.258 { 00:07:52.258 "job": "raid_bdev1", 00:07:52.258 "core_mask": "0x1", 00:07:52.258 "workload": "randrw", 00:07:52.258 "percentage": 50, 00:07:52.258 "status": "finished", 00:07:52.259 "queue_depth": 1, 00:07:52.259 "io_size": 131072, 00:07:52.259 "runtime": 1.381699, 00:07:52.259 "iops": 23698.359773004107, 00:07:52.259 "mibps": 2962.2949716255134, 00:07:52.259 "io_failed": 0, 00:07:52.259 "io_timeout": 0, 00:07:52.259 "avg_latency_us": 39.71521342754751, 00:07:52.259 "min_latency_us": 21.128384279475984, 00:07:52.259 "max_latency_us": 1352.216593886463 00:07:52.259 } 00:07:52.259 ], 00:07:52.259 "core_count": 1 00:07:52.259 } 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74606 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 74606 ']' 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 74606 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74606 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74606' 00:07:52.259 killing process with pid 74606 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 74606 00:07:52.259 [2024-09-29 16:46:13.917519] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:52.259 16:46:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 74606 00:07:52.517 [2024-09-29 16:46:13.932709] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:52.517 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.l0Dt80eNQN 00:07:52.517 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:52.517 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:52.518 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:07:52.518 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:07:52.518 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:52.518 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:07:52.518 16:46:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:07:52.518 00:07:52.518 real 0m3.234s 00:07:52.518 user 0m4.092s 00:07:52.518 sys 0m0.517s 00:07:52.518 16:46:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:52.518 ************************************ 00:07:52.518 END TEST raid_write_error_test 00:07:52.518 ************************************ 00:07:52.518 16:46:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.779 16:46:14 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:52.779 16:46:14 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:52.779 16:46:14 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:07:52.779 16:46:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:52.779 16:46:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:52.779 16:46:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:52.779 ************************************ 00:07:52.779 START TEST raid_state_function_test 00:07:52.779 ************************************ 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 false 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74733 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74733' 00:07:52.779 Process raid pid: 74733 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74733 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 74733 ']' 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:52.779 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:52.779 16:46:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:52.779 [2024-09-29 16:46:14.335275] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:07:52.779 [2024-09-29 16:46:14.335497] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:53.038 [2024-09-29 16:46:14.481509] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:53.038 [2024-09-29 16:46:14.527340] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:53.038 [2024-09-29 16:46:14.569230] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:53.038 [2024-09-29 16:46:14.569348] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.605 [2024-09-29 16:46:15.158274] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:53.605 [2024-09-29 16:46:15.158327] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:53.605 [2024-09-29 16:46:15.158338] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:53.605 [2024-09-29 16:46:15.158349] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:53.605 [2024-09-29 16:46:15.158355] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:53.605 [2024-09-29 16:46:15.158367] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.605 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.605 "name": "Existed_Raid", 00:07:53.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.605 "strip_size_kb": 64, 00:07:53.605 "state": "configuring", 00:07:53.605 "raid_level": "raid0", 00:07:53.605 "superblock": false, 00:07:53.605 "num_base_bdevs": 3, 00:07:53.605 "num_base_bdevs_discovered": 0, 00:07:53.605 "num_base_bdevs_operational": 3, 00:07:53.605 "base_bdevs_list": [ 00:07:53.605 { 00:07:53.605 "name": "BaseBdev1", 00:07:53.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.605 "is_configured": false, 00:07:53.605 "data_offset": 0, 00:07:53.605 "data_size": 0 00:07:53.605 }, 00:07:53.605 { 00:07:53.605 "name": "BaseBdev2", 00:07:53.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.605 "is_configured": false, 00:07:53.605 "data_offset": 0, 00:07:53.605 "data_size": 0 00:07:53.605 }, 00:07:53.605 { 00:07:53.605 "name": "BaseBdev3", 00:07:53.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.605 "is_configured": false, 00:07:53.605 "data_offset": 0, 00:07:53.605 "data_size": 0 00:07:53.605 } 00:07:53.605 ] 00:07:53.605 }' 00:07:53.606 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.606 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.183 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:54.183 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.183 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.183 [2024-09-29 16:46:15.621413] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:54.183 [2024-09-29 16:46:15.621498] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.184 [2024-09-29 16:46:15.633406] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:54.184 [2024-09-29 16:46:15.633489] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:54.184 [2024-09-29 16:46:15.633517] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:54.184 [2024-09-29 16:46:15.633540] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:54.184 [2024-09-29 16:46:15.633557] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:54.184 [2024-09-29 16:46:15.633577] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.184 [2024-09-29 16:46:15.654309] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:54.184 BaseBdev1 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.184 [ 00:07:54.184 { 00:07:54.184 "name": "BaseBdev1", 00:07:54.184 "aliases": [ 00:07:54.184 "4d342cdd-3cc1-4c16-913f-5d1975ee32a0" 00:07:54.184 ], 00:07:54.184 "product_name": "Malloc disk", 00:07:54.184 "block_size": 512, 00:07:54.184 "num_blocks": 65536, 00:07:54.184 "uuid": "4d342cdd-3cc1-4c16-913f-5d1975ee32a0", 00:07:54.184 "assigned_rate_limits": { 00:07:54.184 "rw_ios_per_sec": 0, 00:07:54.184 "rw_mbytes_per_sec": 0, 00:07:54.184 "r_mbytes_per_sec": 0, 00:07:54.184 "w_mbytes_per_sec": 0 00:07:54.184 }, 00:07:54.184 "claimed": true, 00:07:54.184 "claim_type": "exclusive_write", 00:07:54.184 "zoned": false, 00:07:54.184 "supported_io_types": { 00:07:54.184 "read": true, 00:07:54.184 "write": true, 00:07:54.184 "unmap": true, 00:07:54.184 "flush": true, 00:07:54.184 "reset": true, 00:07:54.184 "nvme_admin": false, 00:07:54.184 "nvme_io": false, 00:07:54.184 "nvme_io_md": false, 00:07:54.184 "write_zeroes": true, 00:07:54.184 "zcopy": true, 00:07:54.184 "get_zone_info": false, 00:07:54.184 "zone_management": false, 00:07:54.184 "zone_append": false, 00:07:54.184 "compare": false, 00:07:54.184 "compare_and_write": false, 00:07:54.184 "abort": true, 00:07:54.184 "seek_hole": false, 00:07:54.184 "seek_data": false, 00:07:54.184 "copy": true, 00:07:54.184 "nvme_iov_md": false 00:07:54.184 }, 00:07:54.184 "memory_domains": [ 00:07:54.184 { 00:07:54.184 "dma_device_id": "system", 00:07:54.184 "dma_device_type": 1 00:07:54.184 }, 00:07:54.184 { 00:07:54.184 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.184 "dma_device_type": 2 00:07:54.184 } 00:07:54.184 ], 00:07:54.184 "driver_specific": {} 00:07:54.184 } 00:07:54.184 ] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.184 "name": "Existed_Raid", 00:07:54.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.184 "strip_size_kb": 64, 00:07:54.184 "state": "configuring", 00:07:54.184 "raid_level": "raid0", 00:07:54.184 "superblock": false, 00:07:54.184 "num_base_bdevs": 3, 00:07:54.184 "num_base_bdevs_discovered": 1, 00:07:54.184 "num_base_bdevs_operational": 3, 00:07:54.184 "base_bdevs_list": [ 00:07:54.184 { 00:07:54.184 "name": "BaseBdev1", 00:07:54.184 "uuid": "4d342cdd-3cc1-4c16-913f-5d1975ee32a0", 00:07:54.184 "is_configured": true, 00:07:54.184 "data_offset": 0, 00:07:54.184 "data_size": 65536 00:07:54.184 }, 00:07:54.184 { 00:07:54.184 "name": "BaseBdev2", 00:07:54.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.184 "is_configured": false, 00:07:54.184 "data_offset": 0, 00:07:54.184 "data_size": 0 00:07:54.184 }, 00:07:54.184 { 00:07:54.184 "name": "BaseBdev3", 00:07:54.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.184 "is_configured": false, 00:07:54.184 "data_offset": 0, 00:07:54.184 "data_size": 0 00:07:54.184 } 00:07:54.184 ] 00:07:54.184 }' 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.184 16:46:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.444 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:54.444 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.444 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.703 [2024-09-29 16:46:16.121546] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:54.703 [2024-09-29 16:46:16.121597] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.703 [2024-09-29 16:46:16.133575] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:54.703 [2024-09-29 16:46:16.135456] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:54.703 [2024-09-29 16:46:16.135498] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:54.703 [2024-09-29 16:46:16.135508] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:07:54.703 [2024-09-29 16:46:16.135518] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:54.703 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:54.704 "name": "Existed_Raid", 00:07:54.704 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.704 "strip_size_kb": 64, 00:07:54.704 "state": "configuring", 00:07:54.704 "raid_level": "raid0", 00:07:54.704 "superblock": false, 00:07:54.704 "num_base_bdevs": 3, 00:07:54.704 "num_base_bdevs_discovered": 1, 00:07:54.704 "num_base_bdevs_operational": 3, 00:07:54.704 "base_bdevs_list": [ 00:07:54.704 { 00:07:54.704 "name": "BaseBdev1", 00:07:54.704 "uuid": "4d342cdd-3cc1-4c16-913f-5d1975ee32a0", 00:07:54.704 "is_configured": true, 00:07:54.704 "data_offset": 0, 00:07:54.704 "data_size": 65536 00:07:54.704 }, 00:07:54.704 { 00:07:54.704 "name": "BaseBdev2", 00:07:54.704 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.704 "is_configured": false, 00:07:54.704 "data_offset": 0, 00:07:54.704 "data_size": 0 00:07:54.704 }, 00:07:54.704 { 00:07:54.704 "name": "BaseBdev3", 00:07:54.704 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:54.704 "is_configured": false, 00:07:54.704 "data_offset": 0, 00:07:54.704 "data_size": 0 00:07:54.704 } 00:07:54.704 ] 00:07:54.704 }' 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:54.704 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.964 [2024-09-29 16:46:16.576349] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:54.964 BaseBdev2 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.964 [ 00:07:54.964 { 00:07:54.964 "name": "BaseBdev2", 00:07:54.964 "aliases": [ 00:07:54.964 "fa06a7b3-4e78-4df2-89cb-eb3f06537fff" 00:07:54.964 ], 00:07:54.964 "product_name": "Malloc disk", 00:07:54.964 "block_size": 512, 00:07:54.964 "num_blocks": 65536, 00:07:54.964 "uuid": "fa06a7b3-4e78-4df2-89cb-eb3f06537fff", 00:07:54.964 "assigned_rate_limits": { 00:07:54.964 "rw_ios_per_sec": 0, 00:07:54.964 "rw_mbytes_per_sec": 0, 00:07:54.964 "r_mbytes_per_sec": 0, 00:07:54.964 "w_mbytes_per_sec": 0 00:07:54.964 }, 00:07:54.964 "claimed": true, 00:07:54.964 "claim_type": "exclusive_write", 00:07:54.964 "zoned": false, 00:07:54.964 "supported_io_types": { 00:07:54.964 "read": true, 00:07:54.964 "write": true, 00:07:54.964 "unmap": true, 00:07:54.964 "flush": true, 00:07:54.964 "reset": true, 00:07:54.964 "nvme_admin": false, 00:07:54.964 "nvme_io": false, 00:07:54.964 "nvme_io_md": false, 00:07:54.964 "write_zeroes": true, 00:07:54.964 "zcopy": true, 00:07:54.964 "get_zone_info": false, 00:07:54.964 "zone_management": false, 00:07:54.964 "zone_append": false, 00:07:54.964 "compare": false, 00:07:54.964 "compare_and_write": false, 00:07:54.964 "abort": true, 00:07:54.964 "seek_hole": false, 00:07:54.964 "seek_data": false, 00:07:54.964 "copy": true, 00:07:54.964 "nvme_iov_md": false 00:07:54.964 }, 00:07:54.964 "memory_domains": [ 00:07:54.964 { 00:07:54.964 "dma_device_id": "system", 00:07:54.964 "dma_device_type": 1 00:07:54.964 }, 00:07:54.964 { 00:07:54.964 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:54.964 "dma_device_type": 2 00:07:54.964 } 00:07:54.964 ], 00:07:54.964 "driver_specific": {} 00:07:54.964 } 00:07:54.964 ] 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.964 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.224 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.224 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.224 "name": "Existed_Raid", 00:07:55.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.224 "strip_size_kb": 64, 00:07:55.224 "state": "configuring", 00:07:55.224 "raid_level": "raid0", 00:07:55.224 "superblock": false, 00:07:55.224 "num_base_bdevs": 3, 00:07:55.224 "num_base_bdevs_discovered": 2, 00:07:55.224 "num_base_bdevs_operational": 3, 00:07:55.224 "base_bdevs_list": [ 00:07:55.224 { 00:07:55.224 "name": "BaseBdev1", 00:07:55.224 "uuid": "4d342cdd-3cc1-4c16-913f-5d1975ee32a0", 00:07:55.224 "is_configured": true, 00:07:55.224 "data_offset": 0, 00:07:55.224 "data_size": 65536 00:07:55.224 }, 00:07:55.224 { 00:07:55.224 "name": "BaseBdev2", 00:07:55.224 "uuid": "fa06a7b3-4e78-4df2-89cb-eb3f06537fff", 00:07:55.224 "is_configured": true, 00:07:55.224 "data_offset": 0, 00:07:55.224 "data_size": 65536 00:07:55.224 }, 00:07:55.224 { 00:07:55.224 "name": "BaseBdev3", 00:07:55.224 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:55.224 "is_configured": false, 00:07:55.224 "data_offset": 0, 00:07:55.224 "data_size": 0 00:07:55.224 } 00:07:55.224 ] 00:07:55.224 }' 00:07:55.224 16:46:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.224 16:46:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.484 [2024-09-29 16:46:17.030491] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:55.484 [2024-09-29 16:46:17.030529] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:55.484 [2024-09-29 16:46:17.030542] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:55.484 [2024-09-29 16:46:17.030845] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:55.484 [2024-09-29 16:46:17.030994] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:55.484 [2024-09-29 16:46:17.031009] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:55.484 [2024-09-29 16:46:17.031242] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:55.484 BaseBdev3 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.484 [ 00:07:55.484 { 00:07:55.484 "name": "BaseBdev3", 00:07:55.484 "aliases": [ 00:07:55.484 "8ea6d788-50e9-4164-9287-d5b965533ace" 00:07:55.484 ], 00:07:55.484 "product_name": "Malloc disk", 00:07:55.484 "block_size": 512, 00:07:55.484 "num_blocks": 65536, 00:07:55.484 "uuid": "8ea6d788-50e9-4164-9287-d5b965533ace", 00:07:55.484 "assigned_rate_limits": { 00:07:55.484 "rw_ios_per_sec": 0, 00:07:55.484 "rw_mbytes_per_sec": 0, 00:07:55.484 "r_mbytes_per_sec": 0, 00:07:55.484 "w_mbytes_per_sec": 0 00:07:55.484 }, 00:07:55.484 "claimed": true, 00:07:55.484 "claim_type": "exclusive_write", 00:07:55.484 "zoned": false, 00:07:55.484 "supported_io_types": { 00:07:55.484 "read": true, 00:07:55.484 "write": true, 00:07:55.484 "unmap": true, 00:07:55.484 "flush": true, 00:07:55.484 "reset": true, 00:07:55.484 "nvme_admin": false, 00:07:55.484 "nvme_io": false, 00:07:55.484 "nvme_io_md": false, 00:07:55.484 "write_zeroes": true, 00:07:55.484 "zcopy": true, 00:07:55.484 "get_zone_info": false, 00:07:55.484 "zone_management": false, 00:07:55.484 "zone_append": false, 00:07:55.484 "compare": false, 00:07:55.484 "compare_and_write": false, 00:07:55.484 "abort": true, 00:07:55.484 "seek_hole": false, 00:07:55.484 "seek_data": false, 00:07:55.484 "copy": true, 00:07:55.484 "nvme_iov_md": false 00:07:55.484 }, 00:07:55.484 "memory_domains": [ 00:07:55.484 { 00:07:55.484 "dma_device_id": "system", 00:07:55.484 "dma_device_type": 1 00:07:55.484 }, 00:07:55.484 { 00:07:55.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:55.484 "dma_device_type": 2 00:07:55.484 } 00:07:55.484 ], 00:07:55.484 "driver_specific": {} 00:07:55.484 } 00:07:55.484 ] 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.484 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:55.484 "name": "Existed_Raid", 00:07:55.484 "uuid": "5e127dc6-a854-4c5a-b171-489bcda25e11", 00:07:55.484 "strip_size_kb": 64, 00:07:55.484 "state": "online", 00:07:55.484 "raid_level": "raid0", 00:07:55.484 "superblock": false, 00:07:55.484 "num_base_bdevs": 3, 00:07:55.484 "num_base_bdevs_discovered": 3, 00:07:55.484 "num_base_bdevs_operational": 3, 00:07:55.484 "base_bdevs_list": [ 00:07:55.484 { 00:07:55.484 "name": "BaseBdev1", 00:07:55.484 "uuid": "4d342cdd-3cc1-4c16-913f-5d1975ee32a0", 00:07:55.484 "is_configured": true, 00:07:55.484 "data_offset": 0, 00:07:55.484 "data_size": 65536 00:07:55.484 }, 00:07:55.484 { 00:07:55.484 "name": "BaseBdev2", 00:07:55.484 "uuid": "fa06a7b3-4e78-4df2-89cb-eb3f06537fff", 00:07:55.484 "is_configured": true, 00:07:55.484 "data_offset": 0, 00:07:55.484 "data_size": 65536 00:07:55.484 }, 00:07:55.485 { 00:07:55.485 "name": "BaseBdev3", 00:07:55.485 "uuid": "8ea6d788-50e9-4164-9287-d5b965533ace", 00:07:55.485 "is_configured": true, 00:07:55.485 "data_offset": 0, 00:07:55.485 "data_size": 65536 00:07:55.485 } 00:07:55.485 ] 00:07:55.485 }' 00:07:55.485 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:55.485 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.054 [2024-09-29 16:46:17.510025] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.054 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:56.054 "name": "Existed_Raid", 00:07:56.054 "aliases": [ 00:07:56.054 "5e127dc6-a854-4c5a-b171-489bcda25e11" 00:07:56.054 ], 00:07:56.054 "product_name": "Raid Volume", 00:07:56.054 "block_size": 512, 00:07:56.054 "num_blocks": 196608, 00:07:56.054 "uuid": "5e127dc6-a854-4c5a-b171-489bcda25e11", 00:07:56.054 "assigned_rate_limits": { 00:07:56.054 "rw_ios_per_sec": 0, 00:07:56.054 "rw_mbytes_per_sec": 0, 00:07:56.054 "r_mbytes_per_sec": 0, 00:07:56.054 "w_mbytes_per_sec": 0 00:07:56.054 }, 00:07:56.054 "claimed": false, 00:07:56.054 "zoned": false, 00:07:56.054 "supported_io_types": { 00:07:56.054 "read": true, 00:07:56.054 "write": true, 00:07:56.054 "unmap": true, 00:07:56.054 "flush": true, 00:07:56.054 "reset": true, 00:07:56.054 "nvme_admin": false, 00:07:56.054 "nvme_io": false, 00:07:56.054 "nvme_io_md": false, 00:07:56.054 "write_zeroes": true, 00:07:56.054 "zcopy": false, 00:07:56.054 "get_zone_info": false, 00:07:56.054 "zone_management": false, 00:07:56.054 "zone_append": false, 00:07:56.054 "compare": false, 00:07:56.054 "compare_and_write": false, 00:07:56.054 "abort": false, 00:07:56.054 "seek_hole": false, 00:07:56.054 "seek_data": false, 00:07:56.054 "copy": false, 00:07:56.054 "nvme_iov_md": false 00:07:56.054 }, 00:07:56.054 "memory_domains": [ 00:07:56.054 { 00:07:56.054 "dma_device_id": "system", 00:07:56.054 "dma_device_type": 1 00:07:56.054 }, 00:07:56.054 { 00:07:56.054 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.054 "dma_device_type": 2 00:07:56.054 }, 00:07:56.054 { 00:07:56.054 "dma_device_id": "system", 00:07:56.054 "dma_device_type": 1 00:07:56.054 }, 00:07:56.054 { 00:07:56.054 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.054 "dma_device_type": 2 00:07:56.054 }, 00:07:56.054 { 00:07:56.054 "dma_device_id": "system", 00:07:56.054 "dma_device_type": 1 00:07:56.054 }, 00:07:56.054 { 00:07:56.054 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.054 "dma_device_type": 2 00:07:56.054 } 00:07:56.054 ], 00:07:56.054 "driver_specific": { 00:07:56.054 "raid": { 00:07:56.054 "uuid": "5e127dc6-a854-4c5a-b171-489bcda25e11", 00:07:56.054 "strip_size_kb": 64, 00:07:56.054 "state": "online", 00:07:56.054 "raid_level": "raid0", 00:07:56.054 "superblock": false, 00:07:56.054 "num_base_bdevs": 3, 00:07:56.054 "num_base_bdevs_discovered": 3, 00:07:56.054 "num_base_bdevs_operational": 3, 00:07:56.054 "base_bdevs_list": [ 00:07:56.054 { 00:07:56.054 "name": "BaseBdev1", 00:07:56.055 "uuid": "4d342cdd-3cc1-4c16-913f-5d1975ee32a0", 00:07:56.055 "is_configured": true, 00:07:56.055 "data_offset": 0, 00:07:56.055 "data_size": 65536 00:07:56.055 }, 00:07:56.055 { 00:07:56.055 "name": "BaseBdev2", 00:07:56.055 "uuid": "fa06a7b3-4e78-4df2-89cb-eb3f06537fff", 00:07:56.055 "is_configured": true, 00:07:56.055 "data_offset": 0, 00:07:56.055 "data_size": 65536 00:07:56.055 }, 00:07:56.055 { 00:07:56.055 "name": "BaseBdev3", 00:07:56.055 "uuid": "8ea6d788-50e9-4164-9287-d5b965533ace", 00:07:56.055 "is_configured": true, 00:07:56.055 "data_offset": 0, 00:07:56.055 "data_size": 65536 00:07:56.055 } 00:07:56.055 ] 00:07:56.055 } 00:07:56.055 } 00:07:56.055 }' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:56.055 BaseBdev2 00:07:56.055 BaseBdev3' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.055 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.314 [2024-09-29 16:46:17.765315] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:56.314 [2024-09-29 16:46:17.765340] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:56.314 [2024-09-29 16:46:17.765388] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:56.314 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.315 "name": "Existed_Raid", 00:07:56.315 "uuid": "5e127dc6-a854-4c5a-b171-489bcda25e11", 00:07:56.315 "strip_size_kb": 64, 00:07:56.315 "state": "offline", 00:07:56.315 "raid_level": "raid0", 00:07:56.315 "superblock": false, 00:07:56.315 "num_base_bdevs": 3, 00:07:56.315 "num_base_bdevs_discovered": 2, 00:07:56.315 "num_base_bdevs_operational": 2, 00:07:56.315 "base_bdevs_list": [ 00:07:56.315 { 00:07:56.315 "name": null, 00:07:56.315 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:56.315 "is_configured": false, 00:07:56.315 "data_offset": 0, 00:07:56.315 "data_size": 65536 00:07:56.315 }, 00:07:56.315 { 00:07:56.315 "name": "BaseBdev2", 00:07:56.315 "uuid": "fa06a7b3-4e78-4df2-89cb-eb3f06537fff", 00:07:56.315 "is_configured": true, 00:07:56.315 "data_offset": 0, 00:07:56.315 "data_size": 65536 00:07:56.315 }, 00:07:56.315 { 00:07:56.315 "name": "BaseBdev3", 00:07:56.315 "uuid": "8ea6d788-50e9-4164-9287-d5b965533ace", 00:07:56.315 "is_configured": true, 00:07:56.315 "data_offset": 0, 00:07:56.315 "data_size": 65536 00:07:56.315 } 00:07:56.315 ] 00:07:56.315 }' 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.315 16:46:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.574 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.833 [2024-09-29 16:46:18.247825] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.833 [2024-09-29 16:46:18.318854] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:56.833 [2024-09-29 16:46:18.318896] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.833 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.834 BaseBdev2 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.834 [ 00:07:56.834 { 00:07:56.834 "name": "BaseBdev2", 00:07:56.834 "aliases": [ 00:07:56.834 "8fafce9a-f65e-4789-b8ea-7ccbfe407a69" 00:07:56.834 ], 00:07:56.834 "product_name": "Malloc disk", 00:07:56.834 "block_size": 512, 00:07:56.834 "num_blocks": 65536, 00:07:56.834 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:56.834 "assigned_rate_limits": { 00:07:56.834 "rw_ios_per_sec": 0, 00:07:56.834 "rw_mbytes_per_sec": 0, 00:07:56.834 "r_mbytes_per_sec": 0, 00:07:56.834 "w_mbytes_per_sec": 0 00:07:56.834 }, 00:07:56.834 "claimed": false, 00:07:56.834 "zoned": false, 00:07:56.834 "supported_io_types": { 00:07:56.834 "read": true, 00:07:56.834 "write": true, 00:07:56.834 "unmap": true, 00:07:56.834 "flush": true, 00:07:56.834 "reset": true, 00:07:56.834 "nvme_admin": false, 00:07:56.834 "nvme_io": false, 00:07:56.834 "nvme_io_md": false, 00:07:56.834 "write_zeroes": true, 00:07:56.834 "zcopy": true, 00:07:56.834 "get_zone_info": false, 00:07:56.834 "zone_management": false, 00:07:56.834 "zone_append": false, 00:07:56.834 "compare": false, 00:07:56.834 "compare_and_write": false, 00:07:56.834 "abort": true, 00:07:56.834 "seek_hole": false, 00:07:56.834 "seek_data": false, 00:07:56.834 "copy": true, 00:07:56.834 "nvme_iov_md": false 00:07:56.834 }, 00:07:56.834 "memory_domains": [ 00:07:56.834 { 00:07:56.834 "dma_device_id": "system", 00:07:56.834 "dma_device_type": 1 00:07:56.834 }, 00:07:56.834 { 00:07:56.834 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.834 "dma_device_type": 2 00:07:56.834 } 00:07:56.834 ], 00:07:56.834 "driver_specific": {} 00:07:56.834 } 00:07:56.834 ] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.834 BaseBdev3 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.834 [ 00:07:56.834 { 00:07:56.834 "name": "BaseBdev3", 00:07:56.834 "aliases": [ 00:07:56.834 "85aabd58-175d-4167-9375-4045dffca4b9" 00:07:56.834 ], 00:07:56.834 "product_name": "Malloc disk", 00:07:56.834 "block_size": 512, 00:07:56.834 "num_blocks": 65536, 00:07:56.834 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:56.834 "assigned_rate_limits": { 00:07:56.834 "rw_ios_per_sec": 0, 00:07:56.834 "rw_mbytes_per_sec": 0, 00:07:56.834 "r_mbytes_per_sec": 0, 00:07:56.834 "w_mbytes_per_sec": 0 00:07:56.834 }, 00:07:56.834 "claimed": false, 00:07:56.834 "zoned": false, 00:07:56.834 "supported_io_types": { 00:07:56.834 "read": true, 00:07:56.834 "write": true, 00:07:56.834 "unmap": true, 00:07:56.834 "flush": true, 00:07:56.834 "reset": true, 00:07:56.834 "nvme_admin": false, 00:07:56.834 "nvme_io": false, 00:07:56.834 "nvme_io_md": false, 00:07:56.834 "write_zeroes": true, 00:07:56.834 "zcopy": true, 00:07:56.834 "get_zone_info": false, 00:07:56.834 "zone_management": false, 00:07:56.834 "zone_append": false, 00:07:56.834 "compare": false, 00:07:56.834 "compare_and_write": false, 00:07:56.834 "abort": true, 00:07:56.834 "seek_hole": false, 00:07:56.834 "seek_data": false, 00:07:56.834 "copy": true, 00:07:56.834 "nvme_iov_md": false 00:07:56.834 }, 00:07:56.834 "memory_domains": [ 00:07:56.834 { 00:07:56.834 "dma_device_id": "system", 00:07:56.834 "dma_device_type": 1 00:07:56.834 }, 00:07:56.834 { 00:07:56.834 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.834 "dma_device_type": 2 00:07:56.834 } 00:07:56.834 ], 00:07:56.834 "driver_specific": {} 00:07:56.834 } 00:07:56.834 ] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.834 [2024-09-29 16:46:18.493347] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:56.834 [2024-09-29 16:46:18.493446] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:56.834 [2024-09-29 16:46:18.493486] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:56.834 [2024-09-29 16:46:18.495347] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.834 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.093 "name": "Existed_Raid", 00:07:57.093 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.093 "strip_size_kb": 64, 00:07:57.093 "state": "configuring", 00:07:57.093 "raid_level": "raid0", 00:07:57.093 "superblock": false, 00:07:57.093 "num_base_bdevs": 3, 00:07:57.093 "num_base_bdevs_discovered": 2, 00:07:57.093 "num_base_bdevs_operational": 3, 00:07:57.093 "base_bdevs_list": [ 00:07:57.093 { 00:07:57.093 "name": "BaseBdev1", 00:07:57.093 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.093 "is_configured": false, 00:07:57.093 "data_offset": 0, 00:07:57.093 "data_size": 0 00:07:57.093 }, 00:07:57.093 { 00:07:57.093 "name": "BaseBdev2", 00:07:57.093 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:57.093 "is_configured": true, 00:07:57.093 "data_offset": 0, 00:07:57.093 "data_size": 65536 00:07:57.093 }, 00:07:57.093 { 00:07:57.093 "name": "BaseBdev3", 00:07:57.093 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:57.093 "is_configured": true, 00:07:57.093 "data_offset": 0, 00:07:57.093 "data_size": 65536 00:07:57.093 } 00:07:57.093 ] 00:07:57.093 }' 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.093 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.352 [2024-09-29 16:46:18.900616] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.352 "name": "Existed_Raid", 00:07:57.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.352 "strip_size_kb": 64, 00:07:57.352 "state": "configuring", 00:07:57.352 "raid_level": "raid0", 00:07:57.352 "superblock": false, 00:07:57.352 "num_base_bdevs": 3, 00:07:57.352 "num_base_bdevs_discovered": 1, 00:07:57.352 "num_base_bdevs_operational": 3, 00:07:57.352 "base_bdevs_list": [ 00:07:57.352 { 00:07:57.352 "name": "BaseBdev1", 00:07:57.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.352 "is_configured": false, 00:07:57.352 "data_offset": 0, 00:07:57.352 "data_size": 0 00:07:57.352 }, 00:07:57.352 { 00:07:57.352 "name": null, 00:07:57.352 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:57.352 "is_configured": false, 00:07:57.352 "data_offset": 0, 00:07:57.352 "data_size": 65536 00:07:57.352 }, 00:07:57.352 { 00:07:57.352 "name": "BaseBdev3", 00:07:57.352 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:57.352 "is_configured": true, 00:07:57.352 "data_offset": 0, 00:07:57.352 "data_size": 65536 00:07:57.352 } 00:07:57.352 ] 00:07:57.352 }' 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.352 16:46:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.918 [2024-09-29 16:46:19.402659] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:57.918 BaseBdev1 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.918 [ 00:07:57.918 { 00:07:57.918 "name": "BaseBdev1", 00:07:57.918 "aliases": [ 00:07:57.918 "6c521fbe-5a03-4f3c-9f81-dad980ae019f" 00:07:57.918 ], 00:07:57.918 "product_name": "Malloc disk", 00:07:57.918 "block_size": 512, 00:07:57.918 "num_blocks": 65536, 00:07:57.918 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:07:57.918 "assigned_rate_limits": { 00:07:57.918 "rw_ios_per_sec": 0, 00:07:57.918 "rw_mbytes_per_sec": 0, 00:07:57.918 "r_mbytes_per_sec": 0, 00:07:57.918 "w_mbytes_per_sec": 0 00:07:57.918 }, 00:07:57.918 "claimed": true, 00:07:57.918 "claim_type": "exclusive_write", 00:07:57.918 "zoned": false, 00:07:57.918 "supported_io_types": { 00:07:57.918 "read": true, 00:07:57.918 "write": true, 00:07:57.918 "unmap": true, 00:07:57.918 "flush": true, 00:07:57.918 "reset": true, 00:07:57.918 "nvme_admin": false, 00:07:57.918 "nvme_io": false, 00:07:57.918 "nvme_io_md": false, 00:07:57.918 "write_zeroes": true, 00:07:57.918 "zcopy": true, 00:07:57.918 "get_zone_info": false, 00:07:57.918 "zone_management": false, 00:07:57.918 "zone_append": false, 00:07:57.918 "compare": false, 00:07:57.918 "compare_and_write": false, 00:07:57.918 "abort": true, 00:07:57.918 "seek_hole": false, 00:07:57.918 "seek_data": false, 00:07:57.918 "copy": true, 00:07:57.918 "nvme_iov_md": false 00:07:57.918 }, 00:07:57.918 "memory_domains": [ 00:07:57.918 { 00:07:57.918 "dma_device_id": "system", 00:07:57.918 "dma_device_type": 1 00:07:57.918 }, 00:07:57.918 { 00:07:57.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.918 "dma_device_type": 2 00:07:57.918 } 00:07:57.918 ], 00:07:57.918 "driver_specific": {} 00:07:57.918 } 00:07:57.918 ] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.918 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.918 "name": "Existed_Raid", 00:07:57.918 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:57.918 "strip_size_kb": 64, 00:07:57.918 "state": "configuring", 00:07:57.918 "raid_level": "raid0", 00:07:57.918 "superblock": false, 00:07:57.918 "num_base_bdevs": 3, 00:07:57.918 "num_base_bdevs_discovered": 2, 00:07:57.918 "num_base_bdevs_operational": 3, 00:07:57.918 "base_bdevs_list": [ 00:07:57.918 { 00:07:57.918 "name": "BaseBdev1", 00:07:57.919 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:07:57.919 "is_configured": true, 00:07:57.919 "data_offset": 0, 00:07:57.919 "data_size": 65536 00:07:57.919 }, 00:07:57.919 { 00:07:57.919 "name": null, 00:07:57.919 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:57.919 "is_configured": false, 00:07:57.919 "data_offset": 0, 00:07:57.919 "data_size": 65536 00:07:57.919 }, 00:07:57.919 { 00:07:57.919 "name": "BaseBdev3", 00:07:57.919 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:57.919 "is_configured": true, 00:07:57.919 "data_offset": 0, 00:07:57.919 "data_size": 65536 00:07:57.919 } 00:07:57.919 ] 00:07:57.919 }' 00:07:57.919 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.919 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.486 [2024-09-29 16:46:19.957785] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:58.486 "name": "Existed_Raid", 00:07:58.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:58.486 "strip_size_kb": 64, 00:07:58.486 "state": "configuring", 00:07:58.486 "raid_level": "raid0", 00:07:58.486 "superblock": false, 00:07:58.486 "num_base_bdevs": 3, 00:07:58.486 "num_base_bdevs_discovered": 1, 00:07:58.486 "num_base_bdevs_operational": 3, 00:07:58.486 "base_bdevs_list": [ 00:07:58.486 { 00:07:58.486 "name": "BaseBdev1", 00:07:58.486 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:07:58.486 "is_configured": true, 00:07:58.486 "data_offset": 0, 00:07:58.486 "data_size": 65536 00:07:58.486 }, 00:07:58.486 { 00:07:58.486 "name": null, 00:07:58.486 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:58.486 "is_configured": false, 00:07:58.486 "data_offset": 0, 00:07:58.486 "data_size": 65536 00:07:58.486 }, 00:07:58.486 { 00:07:58.486 "name": null, 00:07:58.486 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:58.486 "is_configured": false, 00:07:58.486 "data_offset": 0, 00:07:58.486 "data_size": 65536 00:07:58.486 } 00:07:58.486 ] 00:07:58.486 }' 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:58.486 16:46:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.745 [2024-09-29 16:46:20.404993] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:58.745 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.004 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.004 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.004 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.004 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.004 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.004 "name": "Existed_Raid", 00:07:59.004 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.004 "strip_size_kb": 64, 00:07:59.004 "state": "configuring", 00:07:59.004 "raid_level": "raid0", 00:07:59.004 "superblock": false, 00:07:59.004 "num_base_bdevs": 3, 00:07:59.004 "num_base_bdevs_discovered": 2, 00:07:59.004 "num_base_bdevs_operational": 3, 00:07:59.004 "base_bdevs_list": [ 00:07:59.004 { 00:07:59.004 "name": "BaseBdev1", 00:07:59.004 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:07:59.004 "is_configured": true, 00:07:59.004 "data_offset": 0, 00:07:59.004 "data_size": 65536 00:07:59.004 }, 00:07:59.004 { 00:07:59.004 "name": null, 00:07:59.004 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:59.004 "is_configured": false, 00:07:59.004 "data_offset": 0, 00:07:59.004 "data_size": 65536 00:07:59.004 }, 00:07:59.004 { 00:07:59.004 "name": "BaseBdev3", 00:07:59.004 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:59.004 "is_configured": true, 00:07:59.004 "data_offset": 0, 00:07:59.004 "data_size": 65536 00:07:59.004 } 00:07:59.004 ] 00:07:59.004 }' 00:07:59.004 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.004 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.263 [2024-09-29 16:46:20.884252] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.263 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.522 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.522 "name": "Existed_Raid", 00:07:59.522 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.522 "strip_size_kb": 64, 00:07:59.522 "state": "configuring", 00:07:59.522 "raid_level": "raid0", 00:07:59.522 "superblock": false, 00:07:59.522 "num_base_bdevs": 3, 00:07:59.522 "num_base_bdevs_discovered": 1, 00:07:59.522 "num_base_bdevs_operational": 3, 00:07:59.522 "base_bdevs_list": [ 00:07:59.522 { 00:07:59.522 "name": null, 00:07:59.522 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:07:59.522 "is_configured": false, 00:07:59.522 "data_offset": 0, 00:07:59.522 "data_size": 65536 00:07:59.522 }, 00:07:59.522 { 00:07:59.522 "name": null, 00:07:59.522 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:59.522 "is_configured": false, 00:07:59.522 "data_offset": 0, 00:07:59.522 "data_size": 65536 00:07:59.522 }, 00:07:59.522 { 00:07:59.522 "name": "BaseBdev3", 00:07:59.522 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:59.522 "is_configured": true, 00:07:59.522 "data_offset": 0, 00:07:59.522 "data_size": 65536 00:07:59.522 } 00:07:59.522 ] 00:07:59.522 }' 00:07:59.522 16:46:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.522 16:46:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.781 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:07:59.781 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.781 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.781 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.781 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.781 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.782 [2024-09-29 16:46:21.385876] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.782 "name": "Existed_Raid", 00:07:59.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:59.782 "strip_size_kb": 64, 00:07:59.782 "state": "configuring", 00:07:59.782 "raid_level": "raid0", 00:07:59.782 "superblock": false, 00:07:59.782 "num_base_bdevs": 3, 00:07:59.782 "num_base_bdevs_discovered": 2, 00:07:59.782 "num_base_bdevs_operational": 3, 00:07:59.782 "base_bdevs_list": [ 00:07:59.782 { 00:07:59.782 "name": null, 00:07:59.782 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:07:59.782 "is_configured": false, 00:07:59.782 "data_offset": 0, 00:07:59.782 "data_size": 65536 00:07:59.782 }, 00:07:59.782 { 00:07:59.782 "name": "BaseBdev2", 00:07:59.782 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:07:59.782 "is_configured": true, 00:07:59.782 "data_offset": 0, 00:07:59.782 "data_size": 65536 00:07:59.782 }, 00:07:59.782 { 00:07:59.782 "name": "BaseBdev3", 00:07:59.782 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:07:59.782 "is_configured": true, 00:07:59.782 "data_offset": 0, 00:07:59.782 "data_size": 65536 00:07:59.782 } 00:07:59.782 ] 00:07:59.782 }' 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.782 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 6c521fbe-5a03-4f3c-9f81-dad980ae019f 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.350 NewBaseBdev 00:08:00.350 [2024-09-29 16:46:21.907851] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:00.350 [2024-09-29 16:46:21.907889] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:00.350 [2024-09-29 16:46:21.907898] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:00.350 [2024-09-29 16:46:21.908127] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:00.350 [2024-09-29 16:46:21.908246] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:00.350 [2024-09-29 16:46:21.908255] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:00.350 [2024-09-29 16:46:21.908461] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:00.350 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.351 [ 00:08:00.351 { 00:08:00.351 "name": "NewBaseBdev", 00:08:00.351 "aliases": [ 00:08:00.351 "6c521fbe-5a03-4f3c-9f81-dad980ae019f" 00:08:00.351 ], 00:08:00.351 "product_name": "Malloc disk", 00:08:00.351 "block_size": 512, 00:08:00.351 "num_blocks": 65536, 00:08:00.351 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:08:00.351 "assigned_rate_limits": { 00:08:00.351 "rw_ios_per_sec": 0, 00:08:00.351 "rw_mbytes_per_sec": 0, 00:08:00.351 "r_mbytes_per_sec": 0, 00:08:00.351 "w_mbytes_per_sec": 0 00:08:00.351 }, 00:08:00.351 "claimed": true, 00:08:00.351 "claim_type": "exclusive_write", 00:08:00.351 "zoned": false, 00:08:00.351 "supported_io_types": { 00:08:00.351 "read": true, 00:08:00.351 "write": true, 00:08:00.351 "unmap": true, 00:08:00.351 "flush": true, 00:08:00.351 "reset": true, 00:08:00.351 "nvme_admin": false, 00:08:00.351 "nvme_io": false, 00:08:00.351 "nvme_io_md": false, 00:08:00.351 "write_zeroes": true, 00:08:00.351 "zcopy": true, 00:08:00.351 "get_zone_info": false, 00:08:00.351 "zone_management": false, 00:08:00.351 "zone_append": false, 00:08:00.351 "compare": false, 00:08:00.351 "compare_and_write": false, 00:08:00.351 "abort": true, 00:08:00.351 "seek_hole": false, 00:08:00.351 "seek_data": false, 00:08:00.351 "copy": true, 00:08:00.351 "nvme_iov_md": false 00:08:00.351 }, 00:08:00.351 "memory_domains": [ 00:08:00.351 { 00:08:00.351 "dma_device_id": "system", 00:08:00.351 "dma_device_type": 1 00:08:00.351 }, 00:08:00.351 { 00:08:00.351 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.351 "dma_device_type": 2 00:08:00.351 } 00:08:00.351 ], 00:08:00.351 "driver_specific": {} 00:08:00.351 } 00:08:00.351 ] 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.351 "name": "Existed_Raid", 00:08:00.351 "uuid": "c70ce118-7feb-4d17-8c8b-3ccbe9f2fc3c", 00:08:00.351 "strip_size_kb": 64, 00:08:00.351 "state": "online", 00:08:00.351 "raid_level": "raid0", 00:08:00.351 "superblock": false, 00:08:00.351 "num_base_bdevs": 3, 00:08:00.351 "num_base_bdevs_discovered": 3, 00:08:00.351 "num_base_bdevs_operational": 3, 00:08:00.351 "base_bdevs_list": [ 00:08:00.351 { 00:08:00.351 "name": "NewBaseBdev", 00:08:00.351 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:08:00.351 "is_configured": true, 00:08:00.351 "data_offset": 0, 00:08:00.351 "data_size": 65536 00:08:00.351 }, 00:08:00.351 { 00:08:00.351 "name": "BaseBdev2", 00:08:00.351 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:08:00.351 "is_configured": true, 00:08:00.351 "data_offset": 0, 00:08:00.351 "data_size": 65536 00:08:00.351 }, 00:08:00.351 { 00:08:00.351 "name": "BaseBdev3", 00:08:00.351 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:08:00.351 "is_configured": true, 00:08:00.351 "data_offset": 0, 00:08:00.351 "data_size": 65536 00:08:00.351 } 00:08:00.351 ] 00:08:00.351 }' 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.351 16:46:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.919 [2024-09-29 16:46:22.371392] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.919 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:00.919 "name": "Existed_Raid", 00:08:00.919 "aliases": [ 00:08:00.919 "c70ce118-7feb-4d17-8c8b-3ccbe9f2fc3c" 00:08:00.919 ], 00:08:00.919 "product_name": "Raid Volume", 00:08:00.919 "block_size": 512, 00:08:00.919 "num_blocks": 196608, 00:08:00.919 "uuid": "c70ce118-7feb-4d17-8c8b-3ccbe9f2fc3c", 00:08:00.919 "assigned_rate_limits": { 00:08:00.919 "rw_ios_per_sec": 0, 00:08:00.919 "rw_mbytes_per_sec": 0, 00:08:00.919 "r_mbytes_per_sec": 0, 00:08:00.919 "w_mbytes_per_sec": 0 00:08:00.919 }, 00:08:00.919 "claimed": false, 00:08:00.919 "zoned": false, 00:08:00.919 "supported_io_types": { 00:08:00.919 "read": true, 00:08:00.919 "write": true, 00:08:00.919 "unmap": true, 00:08:00.919 "flush": true, 00:08:00.919 "reset": true, 00:08:00.919 "nvme_admin": false, 00:08:00.919 "nvme_io": false, 00:08:00.919 "nvme_io_md": false, 00:08:00.919 "write_zeroes": true, 00:08:00.919 "zcopy": false, 00:08:00.919 "get_zone_info": false, 00:08:00.919 "zone_management": false, 00:08:00.919 "zone_append": false, 00:08:00.919 "compare": false, 00:08:00.919 "compare_and_write": false, 00:08:00.919 "abort": false, 00:08:00.919 "seek_hole": false, 00:08:00.919 "seek_data": false, 00:08:00.919 "copy": false, 00:08:00.919 "nvme_iov_md": false 00:08:00.919 }, 00:08:00.919 "memory_domains": [ 00:08:00.919 { 00:08:00.919 "dma_device_id": "system", 00:08:00.919 "dma_device_type": 1 00:08:00.919 }, 00:08:00.919 { 00:08:00.919 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.919 "dma_device_type": 2 00:08:00.919 }, 00:08:00.919 { 00:08:00.919 "dma_device_id": "system", 00:08:00.919 "dma_device_type": 1 00:08:00.919 }, 00:08:00.919 { 00:08:00.919 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.919 "dma_device_type": 2 00:08:00.919 }, 00:08:00.919 { 00:08:00.919 "dma_device_id": "system", 00:08:00.919 "dma_device_type": 1 00:08:00.919 }, 00:08:00.919 { 00:08:00.919 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:00.919 "dma_device_type": 2 00:08:00.919 } 00:08:00.919 ], 00:08:00.920 "driver_specific": { 00:08:00.920 "raid": { 00:08:00.920 "uuid": "c70ce118-7feb-4d17-8c8b-3ccbe9f2fc3c", 00:08:00.920 "strip_size_kb": 64, 00:08:00.920 "state": "online", 00:08:00.920 "raid_level": "raid0", 00:08:00.920 "superblock": false, 00:08:00.920 "num_base_bdevs": 3, 00:08:00.920 "num_base_bdevs_discovered": 3, 00:08:00.920 "num_base_bdevs_operational": 3, 00:08:00.920 "base_bdevs_list": [ 00:08:00.920 { 00:08:00.920 "name": "NewBaseBdev", 00:08:00.920 "uuid": "6c521fbe-5a03-4f3c-9f81-dad980ae019f", 00:08:00.920 "is_configured": true, 00:08:00.920 "data_offset": 0, 00:08:00.920 "data_size": 65536 00:08:00.920 }, 00:08:00.920 { 00:08:00.920 "name": "BaseBdev2", 00:08:00.920 "uuid": "8fafce9a-f65e-4789-b8ea-7ccbfe407a69", 00:08:00.920 "is_configured": true, 00:08:00.920 "data_offset": 0, 00:08:00.920 "data_size": 65536 00:08:00.920 }, 00:08:00.920 { 00:08:00.920 "name": "BaseBdev3", 00:08:00.920 "uuid": "85aabd58-175d-4167-9375-4045dffca4b9", 00:08:00.920 "is_configured": true, 00:08:00.920 "data_offset": 0, 00:08:00.920 "data_size": 65536 00:08:00.920 } 00:08:00.920 ] 00:08:00.920 } 00:08:00.920 } 00:08:00.920 }' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:00.920 BaseBdev2 00:08:00.920 BaseBdev3' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:00.920 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.179 [2024-09-29 16:46:22.666568] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:01.179 [2024-09-29 16:46:22.666633] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:01.179 [2024-09-29 16:46:22.666705] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:01.179 [2024-09-29 16:46:22.666770] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:01.179 [2024-09-29 16:46:22.666783] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74733 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 74733 ']' 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 74733 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74733 00:08:01.179 killing process with pid 74733 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74733' 00:08:01.179 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 74733 00:08:01.180 [2024-09-29 16:46:22.716562] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:01.180 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 74733 00:08:01.180 [2024-09-29 16:46:22.746413] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:01.439 16:46:22 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:01.439 00:08:01.439 real 0m8.744s 00:08:01.439 user 0m14.932s 00:08:01.439 sys 0m1.720s 00:08:01.439 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:01.439 ************************************ 00:08:01.439 END TEST raid_state_function_test 00:08:01.439 ************************************ 00:08:01.439 16:46:22 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.439 16:46:23 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:08:01.440 16:46:23 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:01.440 16:46:23 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:01.440 16:46:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:01.440 ************************************ 00:08:01.440 START TEST raid_state_function_test_sb 00:08:01.440 ************************************ 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 true 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:01.440 Process raid pid: 75338 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75338 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75338' 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75338 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 75338 ']' 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:01.440 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:01.440 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:01.699 [2024-09-29 16:46:23.154460] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:01.699 [2024-09-29 16:46:23.154649] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:01.699 [2024-09-29 16:46:23.299799] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:01.699 [2024-09-29 16:46:23.345873] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:01.958 [2024-09-29 16:46:23.388097] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:01.959 [2024-09-29 16:46:23.388225] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.529 [2024-09-29 16:46:23.977369] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:02.529 [2024-09-29 16:46:23.977421] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:02.529 [2024-09-29 16:46:23.977441] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:02.529 [2024-09-29 16:46:23.977451] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:02.529 [2024-09-29 16:46:23.977458] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:02.529 [2024-09-29 16:46:23.977469] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.529 16:46:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.529 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.529 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.529 "name": "Existed_Raid", 00:08:02.529 "uuid": "7d7e1b07-12be-4db7-8c0b-a663b0dd6138", 00:08:02.529 "strip_size_kb": 64, 00:08:02.529 "state": "configuring", 00:08:02.529 "raid_level": "raid0", 00:08:02.529 "superblock": true, 00:08:02.529 "num_base_bdevs": 3, 00:08:02.529 "num_base_bdevs_discovered": 0, 00:08:02.529 "num_base_bdevs_operational": 3, 00:08:02.529 "base_bdevs_list": [ 00:08:02.529 { 00:08:02.529 "name": "BaseBdev1", 00:08:02.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:02.529 "is_configured": false, 00:08:02.529 "data_offset": 0, 00:08:02.529 "data_size": 0 00:08:02.529 }, 00:08:02.529 { 00:08:02.529 "name": "BaseBdev2", 00:08:02.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:02.529 "is_configured": false, 00:08:02.529 "data_offset": 0, 00:08:02.529 "data_size": 0 00:08:02.529 }, 00:08:02.529 { 00:08:02.529 "name": "BaseBdev3", 00:08:02.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:02.529 "is_configured": false, 00:08:02.529 "data_offset": 0, 00:08:02.529 "data_size": 0 00:08:02.529 } 00:08:02.529 ] 00:08:02.529 }' 00:08:02.529 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.529 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.789 [2024-09-29 16:46:24.400515] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:02.789 [2024-09-29 16:46:24.400554] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.789 [2024-09-29 16:46:24.412536] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:02.789 [2024-09-29 16:46:24.412627] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:02.789 [2024-09-29 16:46:24.412654] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:02.789 [2024-09-29 16:46:24.412675] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:02.789 [2024-09-29 16:46:24.412693] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:02.789 [2024-09-29 16:46:24.412713] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.789 [2024-09-29 16:46:24.433409] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:02.789 BaseBdev1 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.789 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.790 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:02.790 [ 00:08:02.790 { 00:08:02.790 "name": "BaseBdev1", 00:08:02.790 "aliases": [ 00:08:02.790 "a198e31b-58c3-4027-9931-a08af58f641f" 00:08:02.790 ], 00:08:02.790 "product_name": "Malloc disk", 00:08:02.790 "block_size": 512, 00:08:02.790 "num_blocks": 65536, 00:08:02.790 "uuid": "a198e31b-58c3-4027-9931-a08af58f641f", 00:08:02.790 "assigned_rate_limits": { 00:08:02.790 "rw_ios_per_sec": 0, 00:08:02.790 "rw_mbytes_per_sec": 0, 00:08:02.790 "r_mbytes_per_sec": 0, 00:08:02.790 "w_mbytes_per_sec": 0 00:08:02.790 }, 00:08:02.790 "claimed": true, 00:08:03.049 "claim_type": "exclusive_write", 00:08:03.049 "zoned": false, 00:08:03.049 "supported_io_types": { 00:08:03.049 "read": true, 00:08:03.049 "write": true, 00:08:03.049 "unmap": true, 00:08:03.049 "flush": true, 00:08:03.049 "reset": true, 00:08:03.049 "nvme_admin": false, 00:08:03.049 "nvme_io": false, 00:08:03.049 "nvme_io_md": false, 00:08:03.049 "write_zeroes": true, 00:08:03.049 "zcopy": true, 00:08:03.049 "get_zone_info": false, 00:08:03.049 "zone_management": false, 00:08:03.049 "zone_append": false, 00:08:03.049 "compare": false, 00:08:03.049 "compare_and_write": false, 00:08:03.049 "abort": true, 00:08:03.049 "seek_hole": false, 00:08:03.049 "seek_data": false, 00:08:03.049 "copy": true, 00:08:03.049 "nvme_iov_md": false 00:08:03.049 }, 00:08:03.049 "memory_domains": [ 00:08:03.049 { 00:08:03.049 "dma_device_id": "system", 00:08:03.049 "dma_device_type": 1 00:08:03.049 }, 00:08:03.049 { 00:08:03.049 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.049 "dma_device_type": 2 00:08:03.049 } 00:08:03.049 ], 00:08:03.049 "driver_specific": {} 00:08:03.049 } 00:08:03.049 ] 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.049 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.049 "name": "Existed_Raid", 00:08:03.049 "uuid": "01ce1885-f8cb-40c7-8cc9-e677fa1c7ce0", 00:08:03.049 "strip_size_kb": 64, 00:08:03.049 "state": "configuring", 00:08:03.049 "raid_level": "raid0", 00:08:03.049 "superblock": true, 00:08:03.049 "num_base_bdevs": 3, 00:08:03.049 "num_base_bdevs_discovered": 1, 00:08:03.049 "num_base_bdevs_operational": 3, 00:08:03.049 "base_bdevs_list": [ 00:08:03.049 { 00:08:03.049 "name": "BaseBdev1", 00:08:03.049 "uuid": "a198e31b-58c3-4027-9931-a08af58f641f", 00:08:03.049 "is_configured": true, 00:08:03.049 "data_offset": 2048, 00:08:03.049 "data_size": 63488 00:08:03.049 }, 00:08:03.049 { 00:08:03.049 "name": "BaseBdev2", 00:08:03.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.049 "is_configured": false, 00:08:03.049 "data_offset": 0, 00:08:03.049 "data_size": 0 00:08:03.049 }, 00:08:03.049 { 00:08:03.049 "name": "BaseBdev3", 00:08:03.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.050 "is_configured": false, 00:08:03.050 "data_offset": 0, 00:08:03.050 "data_size": 0 00:08:03.050 } 00:08:03.050 ] 00:08:03.050 }' 00:08:03.050 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.050 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.309 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.310 [2024-09-29 16:46:24.932594] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:03.310 [2024-09-29 16:46:24.932703] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.310 [2024-09-29 16:46:24.944635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:03.310 [2024-09-29 16:46:24.946505] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:03.310 [2024-09-29 16:46:24.946578] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:03.310 [2024-09-29 16:46:24.946605] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:03.310 [2024-09-29 16:46:24.946628] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.310 16:46:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.570 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:03.570 "name": "Existed_Raid", 00:08:03.570 "uuid": "87f79c1f-6ba0-4506-bbf2-017775c265f0", 00:08:03.570 "strip_size_kb": 64, 00:08:03.570 "state": "configuring", 00:08:03.570 "raid_level": "raid0", 00:08:03.570 "superblock": true, 00:08:03.570 "num_base_bdevs": 3, 00:08:03.570 "num_base_bdevs_discovered": 1, 00:08:03.570 "num_base_bdevs_operational": 3, 00:08:03.570 "base_bdevs_list": [ 00:08:03.570 { 00:08:03.570 "name": "BaseBdev1", 00:08:03.570 "uuid": "a198e31b-58c3-4027-9931-a08af58f641f", 00:08:03.570 "is_configured": true, 00:08:03.570 "data_offset": 2048, 00:08:03.570 "data_size": 63488 00:08:03.570 }, 00:08:03.570 { 00:08:03.570 "name": "BaseBdev2", 00:08:03.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.570 "is_configured": false, 00:08:03.570 "data_offset": 0, 00:08:03.570 "data_size": 0 00:08:03.570 }, 00:08:03.570 { 00:08:03.570 "name": "BaseBdev3", 00:08:03.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:03.570 "is_configured": false, 00:08:03.570 "data_offset": 0, 00:08:03.570 "data_size": 0 00:08:03.570 } 00:08:03.570 ] 00:08:03.570 }' 00:08:03.570 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:03.570 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.829 [2024-09-29 16:46:25.418317] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:03.829 BaseBdev2 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.829 [ 00:08:03.829 { 00:08:03.829 "name": "BaseBdev2", 00:08:03.829 "aliases": [ 00:08:03.829 "bde61634-0964-445f-89f9-31bf64b23530" 00:08:03.829 ], 00:08:03.829 "product_name": "Malloc disk", 00:08:03.829 "block_size": 512, 00:08:03.829 "num_blocks": 65536, 00:08:03.829 "uuid": "bde61634-0964-445f-89f9-31bf64b23530", 00:08:03.829 "assigned_rate_limits": { 00:08:03.829 "rw_ios_per_sec": 0, 00:08:03.829 "rw_mbytes_per_sec": 0, 00:08:03.829 "r_mbytes_per_sec": 0, 00:08:03.829 "w_mbytes_per_sec": 0 00:08:03.829 }, 00:08:03.829 "claimed": true, 00:08:03.829 "claim_type": "exclusive_write", 00:08:03.829 "zoned": false, 00:08:03.829 "supported_io_types": { 00:08:03.829 "read": true, 00:08:03.829 "write": true, 00:08:03.829 "unmap": true, 00:08:03.829 "flush": true, 00:08:03.829 "reset": true, 00:08:03.829 "nvme_admin": false, 00:08:03.829 "nvme_io": false, 00:08:03.829 "nvme_io_md": false, 00:08:03.829 "write_zeroes": true, 00:08:03.829 "zcopy": true, 00:08:03.829 "get_zone_info": false, 00:08:03.829 "zone_management": false, 00:08:03.829 "zone_append": false, 00:08:03.829 "compare": false, 00:08:03.829 "compare_and_write": false, 00:08:03.829 "abort": true, 00:08:03.829 "seek_hole": false, 00:08:03.829 "seek_data": false, 00:08:03.829 "copy": true, 00:08:03.829 "nvme_iov_md": false 00:08:03.829 }, 00:08:03.829 "memory_domains": [ 00:08:03.829 { 00:08:03.829 "dma_device_id": "system", 00:08:03.829 "dma_device_type": 1 00:08:03.829 }, 00:08:03.829 { 00:08:03.829 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:03.829 "dma_device_type": 2 00:08:03.829 } 00:08:03.829 ], 00:08:03.829 "driver_specific": {} 00:08:03.829 } 00:08:03.829 ] 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:03.829 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:03.830 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.089 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.089 "name": "Existed_Raid", 00:08:04.089 "uuid": "87f79c1f-6ba0-4506-bbf2-017775c265f0", 00:08:04.089 "strip_size_kb": 64, 00:08:04.089 "state": "configuring", 00:08:04.089 "raid_level": "raid0", 00:08:04.089 "superblock": true, 00:08:04.089 "num_base_bdevs": 3, 00:08:04.089 "num_base_bdevs_discovered": 2, 00:08:04.089 "num_base_bdevs_operational": 3, 00:08:04.089 "base_bdevs_list": [ 00:08:04.089 { 00:08:04.089 "name": "BaseBdev1", 00:08:04.089 "uuid": "a198e31b-58c3-4027-9931-a08af58f641f", 00:08:04.089 "is_configured": true, 00:08:04.089 "data_offset": 2048, 00:08:04.089 "data_size": 63488 00:08:04.089 }, 00:08:04.089 { 00:08:04.089 "name": "BaseBdev2", 00:08:04.089 "uuid": "bde61634-0964-445f-89f9-31bf64b23530", 00:08:04.089 "is_configured": true, 00:08:04.089 "data_offset": 2048, 00:08:04.089 "data_size": 63488 00:08:04.089 }, 00:08:04.089 { 00:08:04.089 "name": "BaseBdev3", 00:08:04.089 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:04.089 "is_configured": false, 00:08:04.089 "data_offset": 0, 00:08:04.089 "data_size": 0 00:08:04.089 } 00:08:04.089 ] 00:08:04.089 }' 00:08:04.089 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.089 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.348 [2024-09-29 16:46:25.900462] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:04.348 [2024-09-29 16:46:25.900651] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:04.348 [2024-09-29 16:46:25.900669] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:04.348 [2024-09-29 16:46:25.900997] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:04.348 BaseBdev3 00:08:04.348 [2024-09-29 16:46:25.901123] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:04.348 [2024-09-29 16:46:25.901133] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:04.348 [2024-09-29 16:46:25.901257] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.348 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.348 [ 00:08:04.348 { 00:08:04.348 "name": "BaseBdev3", 00:08:04.348 "aliases": [ 00:08:04.348 "32017645-639c-4431-9dd0-4a1759af235c" 00:08:04.348 ], 00:08:04.348 "product_name": "Malloc disk", 00:08:04.348 "block_size": 512, 00:08:04.348 "num_blocks": 65536, 00:08:04.348 "uuid": "32017645-639c-4431-9dd0-4a1759af235c", 00:08:04.348 "assigned_rate_limits": { 00:08:04.348 "rw_ios_per_sec": 0, 00:08:04.348 "rw_mbytes_per_sec": 0, 00:08:04.348 "r_mbytes_per_sec": 0, 00:08:04.348 "w_mbytes_per_sec": 0 00:08:04.348 }, 00:08:04.348 "claimed": true, 00:08:04.348 "claim_type": "exclusive_write", 00:08:04.348 "zoned": false, 00:08:04.348 "supported_io_types": { 00:08:04.348 "read": true, 00:08:04.348 "write": true, 00:08:04.348 "unmap": true, 00:08:04.348 "flush": true, 00:08:04.348 "reset": true, 00:08:04.348 "nvme_admin": false, 00:08:04.349 "nvme_io": false, 00:08:04.349 "nvme_io_md": false, 00:08:04.349 "write_zeroes": true, 00:08:04.349 "zcopy": true, 00:08:04.349 "get_zone_info": false, 00:08:04.349 "zone_management": false, 00:08:04.349 "zone_append": false, 00:08:04.349 "compare": false, 00:08:04.349 "compare_and_write": false, 00:08:04.349 "abort": true, 00:08:04.349 "seek_hole": false, 00:08:04.349 "seek_data": false, 00:08:04.349 "copy": true, 00:08:04.349 "nvme_iov_md": false 00:08:04.349 }, 00:08:04.349 "memory_domains": [ 00:08:04.349 { 00:08:04.349 "dma_device_id": "system", 00:08:04.349 "dma_device_type": 1 00:08:04.349 }, 00:08:04.349 { 00:08:04.349 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.349 "dma_device_type": 2 00:08:04.349 } 00:08:04.349 ], 00:08:04.349 "driver_specific": {} 00:08:04.349 } 00:08:04.349 ] 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.349 "name": "Existed_Raid", 00:08:04.349 "uuid": "87f79c1f-6ba0-4506-bbf2-017775c265f0", 00:08:04.349 "strip_size_kb": 64, 00:08:04.349 "state": "online", 00:08:04.349 "raid_level": "raid0", 00:08:04.349 "superblock": true, 00:08:04.349 "num_base_bdevs": 3, 00:08:04.349 "num_base_bdevs_discovered": 3, 00:08:04.349 "num_base_bdevs_operational": 3, 00:08:04.349 "base_bdevs_list": [ 00:08:04.349 { 00:08:04.349 "name": "BaseBdev1", 00:08:04.349 "uuid": "a198e31b-58c3-4027-9931-a08af58f641f", 00:08:04.349 "is_configured": true, 00:08:04.349 "data_offset": 2048, 00:08:04.349 "data_size": 63488 00:08:04.349 }, 00:08:04.349 { 00:08:04.349 "name": "BaseBdev2", 00:08:04.349 "uuid": "bde61634-0964-445f-89f9-31bf64b23530", 00:08:04.349 "is_configured": true, 00:08:04.349 "data_offset": 2048, 00:08:04.349 "data_size": 63488 00:08:04.349 }, 00:08:04.349 { 00:08:04.349 "name": "BaseBdev3", 00:08:04.349 "uuid": "32017645-639c-4431-9dd0-4a1759af235c", 00:08:04.349 "is_configured": true, 00:08:04.349 "data_offset": 2048, 00:08:04.349 "data_size": 63488 00:08:04.349 } 00:08:04.349 ] 00:08:04.349 }' 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.349 16:46:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.918 [2024-09-29 16:46:26.435871] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:04.918 "name": "Existed_Raid", 00:08:04.918 "aliases": [ 00:08:04.918 "87f79c1f-6ba0-4506-bbf2-017775c265f0" 00:08:04.918 ], 00:08:04.918 "product_name": "Raid Volume", 00:08:04.918 "block_size": 512, 00:08:04.918 "num_blocks": 190464, 00:08:04.918 "uuid": "87f79c1f-6ba0-4506-bbf2-017775c265f0", 00:08:04.918 "assigned_rate_limits": { 00:08:04.918 "rw_ios_per_sec": 0, 00:08:04.918 "rw_mbytes_per_sec": 0, 00:08:04.918 "r_mbytes_per_sec": 0, 00:08:04.918 "w_mbytes_per_sec": 0 00:08:04.918 }, 00:08:04.918 "claimed": false, 00:08:04.918 "zoned": false, 00:08:04.918 "supported_io_types": { 00:08:04.918 "read": true, 00:08:04.918 "write": true, 00:08:04.918 "unmap": true, 00:08:04.918 "flush": true, 00:08:04.918 "reset": true, 00:08:04.918 "nvme_admin": false, 00:08:04.918 "nvme_io": false, 00:08:04.918 "nvme_io_md": false, 00:08:04.918 "write_zeroes": true, 00:08:04.918 "zcopy": false, 00:08:04.918 "get_zone_info": false, 00:08:04.918 "zone_management": false, 00:08:04.918 "zone_append": false, 00:08:04.918 "compare": false, 00:08:04.918 "compare_and_write": false, 00:08:04.918 "abort": false, 00:08:04.918 "seek_hole": false, 00:08:04.918 "seek_data": false, 00:08:04.918 "copy": false, 00:08:04.918 "nvme_iov_md": false 00:08:04.918 }, 00:08:04.918 "memory_domains": [ 00:08:04.918 { 00:08:04.918 "dma_device_id": "system", 00:08:04.918 "dma_device_type": 1 00:08:04.918 }, 00:08:04.918 { 00:08:04.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.918 "dma_device_type": 2 00:08:04.918 }, 00:08:04.918 { 00:08:04.918 "dma_device_id": "system", 00:08:04.918 "dma_device_type": 1 00:08:04.918 }, 00:08:04.918 { 00:08:04.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.918 "dma_device_type": 2 00:08:04.918 }, 00:08:04.918 { 00:08:04.918 "dma_device_id": "system", 00:08:04.918 "dma_device_type": 1 00:08:04.918 }, 00:08:04.918 { 00:08:04.918 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:04.918 "dma_device_type": 2 00:08:04.918 } 00:08:04.918 ], 00:08:04.918 "driver_specific": { 00:08:04.918 "raid": { 00:08:04.918 "uuid": "87f79c1f-6ba0-4506-bbf2-017775c265f0", 00:08:04.918 "strip_size_kb": 64, 00:08:04.918 "state": "online", 00:08:04.918 "raid_level": "raid0", 00:08:04.918 "superblock": true, 00:08:04.918 "num_base_bdevs": 3, 00:08:04.918 "num_base_bdevs_discovered": 3, 00:08:04.918 "num_base_bdevs_operational": 3, 00:08:04.918 "base_bdevs_list": [ 00:08:04.918 { 00:08:04.918 "name": "BaseBdev1", 00:08:04.918 "uuid": "a198e31b-58c3-4027-9931-a08af58f641f", 00:08:04.918 "is_configured": true, 00:08:04.918 "data_offset": 2048, 00:08:04.918 "data_size": 63488 00:08:04.918 }, 00:08:04.918 { 00:08:04.918 "name": "BaseBdev2", 00:08:04.918 "uuid": "bde61634-0964-445f-89f9-31bf64b23530", 00:08:04.918 "is_configured": true, 00:08:04.918 "data_offset": 2048, 00:08:04.918 "data_size": 63488 00:08:04.918 }, 00:08:04.918 { 00:08:04.918 "name": "BaseBdev3", 00:08:04.918 "uuid": "32017645-639c-4431-9dd0-4a1759af235c", 00:08:04.918 "is_configured": true, 00:08:04.918 "data_offset": 2048, 00:08:04.918 "data_size": 63488 00:08:04.918 } 00:08:04.918 ] 00:08:04.918 } 00:08:04.918 } 00:08:04.918 }' 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:04.918 BaseBdev2 00:08:04.918 BaseBdev3' 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:04.918 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.178 [2024-09-29 16:46:26.719172] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:05.178 [2024-09-29 16:46:26.719234] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:05.178 [2024-09-29 16:46:26.719309] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:05.178 "name": "Existed_Raid", 00:08:05.178 "uuid": "87f79c1f-6ba0-4506-bbf2-017775c265f0", 00:08:05.178 "strip_size_kb": 64, 00:08:05.178 "state": "offline", 00:08:05.178 "raid_level": "raid0", 00:08:05.178 "superblock": true, 00:08:05.178 "num_base_bdevs": 3, 00:08:05.178 "num_base_bdevs_discovered": 2, 00:08:05.178 "num_base_bdevs_operational": 2, 00:08:05.178 "base_bdevs_list": [ 00:08:05.178 { 00:08:05.178 "name": null, 00:08:05.178 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:05.178 "is_configured": false, 00:08:05.178 "data_offset": 0, 00:08:05.178 "data_size": 63488 00:08:05.178 }, 00:08:05.178 { 00:08:05.178 "name": "BaseBdev2", 00:08:05.178 "uuid": "bde61634-0964-445f-89f9-31bf64b23530", 00:08:05.178 "is_configured": true, 00:08:05.178 "data_offset": 2048, 00:08:05.178 "data_size": 63488 00:08:05.178 }, 00:08:05.178 { 00:08:05.178 "name": "BaseBdev3", 00:08:05.178 "uuid": "32017645-639c-4431-9dd0-4a1759af235c", 00:08:05.178 "is_configured": true, 00:08:05.178 "data_offset": 2048, 00:08:05.178 "data_size": 63488 00:08:05.178 } 00:08:05.178 ] 00:08:05.178 }' 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:05.178 16:46:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 [2024-09-29 16:46:27.225625] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 [2024-09-29 16:46:27.292649] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:05.747 [2024-09-29 16:46:27.292744] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 BaseBdev2 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:05.747 [ 00:08:05.747 { 00:08:05.747 "name": "BaseBdev2", 00:08:05.747 "aliases": [ 00:08:05.747 "5bf6560d-5d66-4389-9ec7-aed7bc0255b9" 00:08:05.747 ], 00:08:05.747 "product_name": "Malloc disk", 00:08:05.747 "block_size": 512, 00:08:05.747 "num_blocks": 65536, 00:08:05.747 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:05.747 "assigned_rate_limits": { 00:08:05.747 "rw_ios_per_sec": 0, 00:08:05.747 "rw_mbytes_per_sec": 0, 00:08:05.747 "r_mbytes_per_sec": 0, 00:08:05.747 "w_mbytes_per_sec": 0 00:08:05.747 }, 00:08:05.747 "claimed": false, 00:08:05.747 "zoned": false, 00:08:05.747 "supported_io_types": { 00:08:05.747 "read": true, 00:08:05.747 "write": true, 00:08:05.747 "unmap": true, 00:08:05.747 "flush": true, 00:08:05.747 "reset": true, 00:08:05.747 "nvme_admin": false, 00:08:05.747 "nvme_io": false, 00:08:05.747 "nvme_io_md": false, 00:08:05.747 "write_zeroes": true, 00:08:05.747 "zcopy": true, 00:08:05.747 "get_zone_info": false, 00:08:05.747 "zone_management": false, 00:08:05.747 "zone_append": false, 00:08:05.747 "compare": false, 00:08:05.747 "compare_and_write": false, 00:08:05.747 "abort": true, 00:08:05.747 "seek_hole": false, 00:08:05.747 "seek_data": false, 00:08:05.747 "copy": true, 00:08:05.747 "nvme_iov_md": false 00:08:05.747 }, 00:08:05.747 "memory_domains": [ 00:08:05.747 { 00:08:05.747 "dma_device_id": "system", 00:08:05.747 "dma_device_type": 1 00:08:05.747 }, 00:08:05.747 { 00:08:05.747 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:05.747 "dma_device_type": 2 00:08:05.747 } 00:08:05.747 ], 00:08:05.747 "driver_specific": {} 00:08:05.747 } 00:08:05.747 ] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:05.747 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.007 BaseBdev3 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.007 [ 00:08:06.007 { 00:08:06.007 "name": "BaseBdev3", 00:08:06.007 "aliases": [ 00:08:06.007 "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1" 00:08:06.007 ], 00:08:06.007 "product_name": "Malloc disk", 00:08:06.007 "block_size": 512, 00:08:06.007 "num_blocks": 65536, 00:08:06.007 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:06.007 "assigned_rate_limits": { 00:08:06.007 "rw_ios_per_sec": 0, 00:08:06.007 "rw_mbytes_per_sec": 0, 00:08:06.007 "r_mbytes_per_sec": 0, 00:08:06.007 "w_mbytes_per_sec": 0 00:08:06.007 }, 00:08:06.007 "claimed": false, 00:08:06.007 "zoned": false, 00:08:06.007 "supported_io_types": { 00:08:06.007 "read": true, 00:08:06.007 "write": true, 00:08:06.007 "unmap": true, 00:08:06.007 "flush": true, 00:08:06.007 "reset": true, 00:08:06.007 "nvme_admin": false, 00:08:06.007 "nvme_io": false, 00:08:06.007 "nvme_io_md": false, 00:08:06.007 "write_zeroes": true, 00:08:06.007 "zcopy": true, 00:08:06.007 "get_zone_info": false, 00:08:06.007 "zone_management": false, 00:08:06.007 "zone_append": false, 00:08:06.007 "compare": false, 00:08:06.007 "compare_and_write": false, 00:08:06.007 "abort": true, 00:08:06.007 "seek_hole": false, 00:08:06.007 "seek_data": false, 00:08:06.007 "copy": true, 00:08:06.007 "nvme_iov_md": false 00:08:06.007 }, 00:08:06.007 "memory_domains": [ 00:08:06.007 { 00:08:06.007 "dma_device_id": "system", 00:08:06.007 "dma_device_type": 1 00:08:06.007 }, 00:08:06.007 { 00:08:06.007 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.007 "dma_device_type": 2 00:08:06.007 } 00:08:06.007 ], 00:08:06.007 "driver_specific": {} 00:08:06.007 } 00:08:06.007 ] 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.007 [2024-09-29 16:46:27.467275] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:06.007 [2024-09-29 16:46:27.467319] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:06.007 [2024-09-29 16:46:27.467338] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:06.007 [2024-09-29 16:46:27.469119] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:06.007 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.008 "name": "Existed_Raid", 00:08:06.008 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:06.008 "strip_size_kb": 64, 00:08:06.008 "state": "configuring", 00:08:06.008 "raid_level": "raid0", 00:08:06.008 "superblock": true, 00:08:06.008 "num_base_bdevs": 3, 00:08:06.008 "num_base_bdevs_discovered": 2, 00:08:06.008 "num_base_bdevs_operational": 3, 00:08:06.008 "base_bdevs_list": [ 00:08:06.008 { 00:08:06.008 "name": "BaseBdev1", 00:08:06.008 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.008 "is_configured": false, 00:08:06.008 "data_offset": 0, 00:08:06.008 "data_size": 0 00:08:06.008 }, 00:08:06.008 { 00:08:06.008 "name": "BaseBdev2", 00:08:06.008 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:06.008 "is_configured": true, 00:08:06.008 "data_offset": 2048, 00:08:06.008 "data_size": 63488 00:08:06.008 }, 00:08:06.008 { 00:08:06.008 "name": "BaseBdev3", 00:08:06.008 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:06.008 "is_configured": true, 00:08:06.008 "data_offset": 2048, 00:08:06.008 "data_size": 63488 00:08:06.008 } 00:08:06.008 ] 00:08:06.008 }' 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.008 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.268 [2024-09-29 16:46:27.902485] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.268 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.527 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.527 "name": "Existed_Raid", 00:08:06.527 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:06.527 "strip_size_kb": 64, 00:08:06.527 "state": "configuring", 00:08:06.527 "raid_level": "raid0", 00:08:06.527 "superblock": true, 00:08:06.527 "num_base_bdevs": 3, 00:08:06.527 "num_base_bdevs_discovered": 1, 00:08:06.527 "num_base_bdevs_operational": 3, 00:08:06.527 "base_bdevs_list": [ 00:08:06.527 { 00:08:06.527 "name": "BaseBdev1", 00:08:06.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.527 "is_configured": false, 00:08:06.527 "data_offset": 0, 00:08:06.527 "data_size": 0 00:08:06.527 }, 00:08:06.527 { 00:08:06.527 "name": null, 00:08:06.527 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:06.527 "is_configured": false, 00:08:06.527 "data_offset": 0, 00:08:06.527 "data_size": 63488 00:08:06.527 }, 00:08:06.527 { 00:08:06.527 "name": "BaseBdev3", 00:08:06.527 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:06.527 "is_configured": true, 00:08:06.527 "data_offset": 2048, 00:08:06.527 "data_size": 63488 00:08:06.527 } 00:08:06.527 ] 00:08:06.527 }' 00:08:06.527 16:46:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.527 16:46:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.804 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.805 BaseBdev1 00:08:06.805 [2024-09-29 16:46:28.376596] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.805 [ 00:08:06.805 { 00:08:06.805 "name": "BaseBdev1", 00:08:06.805 "aliases": [ 00:08:06.805 "4182a219-21ea-41fa-a691-9c5c7624689c" 00:08:06.805 ], 00:08:06.805 "product_name": "Malloc disk", 00:08:06.805 "block_size": 512, 00:08:06.805 "num_blocks": 65536, 00:08:06.805 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:06.805 "assigned_rate_limits": { 00:08:06.805 "rw_ios_per_sec": 0, 00:08:06.805 "rw_mbytes_per_sec": 0, 00:08:06.805 "r_mbytes_per_sec": 0, 00:08:06.805 "w_mbytes_per_sec": 0 00:08:06.805 }, 00:08:06.805 "claimed": true, 00:08:06.805 "claim_type": "exclusive_write", 00:08:06.805 "zoned": false, 00:08:06.805 "supported_io_types": { 00:08:06.805 "read": true, 00:08:06.805 "write": true, 00:08:06.805 "unmap": true, 00:08:06.805 "flush": true, 00:08:06.805 "reset": true, 00:08:06.805 "nvme_admin": false, 00:08:06.805 "nvme_io": false, 00:08:06.805 "nvme_io_md": false, 00:08:06.805 "write_zeroes": true, 00:08:06.805 "zcopy": true, 00:08:06.805 "get_zone_info": false, 00:08:06.805 "zone_management": false, 00:08:06.805 "zone_append": false, 00:08:06.805 "compare": false, 00:08:06.805 "compare_and_write": false, 00:08:06.805 "abort": true, 00:08:06.805 "seek_hole": false, 00:08:06.805 "seek_data": false, 00:08:06.805 "copy": true, 00:08:06.805 "nvme_iov_md": false 00:08:06.805 }, 00:08:06.805 "memory_domains": [ 00:08:06.805 { 00:08:06.805 "dma_device_id": "system", 00:08:06.805 "dma_device_type": 1 00:08:06.805 }, 00:08:06.805 { 00:08:06.805 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.805 "dma_device_type": 2 00:08:06.805 } 00:08:06.805 ], 00:08:06.805 "driver_specific": {} 00:08:06.805 } 00:08:06.805 ] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.805 "name": "Existed_Raid", 00:08:06.805 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:06.805 "strip_size_kb": 64, 00:08:06.805 "state": "configuring", 00:08:06.805 "raid_level": "raid0", 00:08:06.805 "superblock": true, 00:08:06.805 "num_base_bdevs": 3, 00:08:06.805 "num_base_bdevs_discovered": 2, 00:08:06.805 "num_base_bdevs_operational": 3, 00:08:06.805 "base_bdevs_list": [ 00:08:06.805 { 00:08:06.805 "name": "BaseBdev1", 00:08:06.805 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:06.805 "is_configured": true, 00:08:06.805 "data_offset": 2048, 00:08:06.805 "data_size": 63488 00:08:06.805 }, 00:08:06.805 { 00:08:06.805 "name": null, 00:08:06.805 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:06.805 "is_configured": false, 00:08:06.805 "data_offset": 0, 00:08:06.805 "data_size": 63488 00:08:06.805 }, 00:08:06.805 { 00:08:06.805 "name": "BaseBdev3", 00:08:06.805 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:06.805 "is_configured": true, 00:08:06.805 "data_offset": 2048, 00:08:06.805 "data_size": 63488 00:08:06.805 } 00:08:06.805 ] 00:08:06.805 }' 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.805 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.391 [2024-09-29 16:46:28.891753] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:07.391 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.392 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.392 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.392 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.392 "name": "Existed_Raid", 00:08:07.392 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:07.392 "strip_size_kb": 64, 00:08:07.392 "state": "configuring", 00:08:07.392 "raid_level": "raid0", 00:08:07.392 "superblock": true, 00:08:07.392 "num_base_bdevs": 3, 00:08:07.392 "num_base_bdevs_discovered": 1, 00:08:07.392 "num_base_bdevs_operational": 3, 00:08:07.392 "base_bdevs_list": [ 00:08:07.392 { 00:08:07.392 "name": "BaseBdev1", 00:08:07.392 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:07.392 "is_configured": true, 00:08:07.392 "data_offset": 2048, 00:08:07.392 "data_size": 63488 00:08:07.392 }, 00:08:07.392 { 00:08:07.392 "name": null, 00:08:07.392 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:07.392 "is_configured": false, 00:08:07.392 "data_offset": 0, 00:08:07.392 "data_size": 63488 00:08:07.392 }, 00:08:07.392 { 00:08:07.392 "name": null, 00:08:07.392 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:07.392 "is_configured": false, 00:08:07.392 "data_offset": 0, 00:08:07.392 "data_size": 63488 00:08:07.392 } 00:08:07.392 ] 00:08:07.392 }' 00:08:07.392 16:46:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.392 16:46:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.960 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.960 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.960 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.960 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:07.960 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.960 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.961 [2024-09-29 16:46:29.382905] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.961 "name": "Existed_Raid", 00:08:07.961 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:07.961 "strip_size_kb": 64, 00:08:07.961 "state": "configuring", 00:08:07.961 "raid_level": "raid0", 00:08:07.961 "superblock": true, 00:08:07.961 "num_base_bdevs": 3, 00:08:07.961 "num_base_bdevs_discovered": 2, 00:08:07.961 "num_base_bdevs_operational": 3, 00:08:07.961 "base_bdevs_list": [ 00:08:07.961 { 00:08:07.961 "name": "BaseBdev1", 00:08:07.961 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:07.961 "is_configured": true, 00:08:07.961 "data_offset": 2048, 00:08:07.961 "data_size": 63488 00:08:07.961 }, 00:08:07.961 { 00:08:07.961 "name": null, 00:08:07.961 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:07.961 "is_configured": false, 00:08:07.961 "data_offset": 0, 00:08:07.961 "data_size": 63488 00:08:07.961 }, 00:08:07.961 { 00:08:07.961 "name": "BaseBdev3", 00:08:07.961 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:07.961 "is_configured": true, 00:08:07.961 "data_offset": 2048, 00:08:07.961 "data_size": 63488 00:08:07.961 } 00:08:07.961 ] 00:08:07.961 }' 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.961 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.221 [2024-09-29 16:46:29.838161] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.221 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.481 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.481 "name": "Existed_Raid", 00:08:08.481 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:08.481 "strip_size_kb": 64, 00:08:08.481 "state": "configuring", 00:08:08.481 "raid_level": "raid0", 00:08:08.481 "superblock": true, 00:08:08.481 "num_base_bdevs": 3, 00:08:08.481 "num_base_bdevs_discovered": 1, 00:08:08.481 "num_base_bdevs_operational": 3, 00:08:08.481 "base_bdevs_list": [ 00:08:08.481 { 00:08:08.481 "name": null, 00:08:08.481 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:08.481 "is_configured": false, 00:08:08.481 "data_offset": 0, 00:08:08.481 "data_size": 63488 00:08:08.481 }, 00:08:08.481 { 00:08:08.481 "name": null, 00:08:08.481 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:08.481 "is_configured": false, 00:08:08.481 "data_offset": 0, 00:08:08.481 "data_size": 63488 00:08:08.481 }, 00:08:08.481 { 00:08:08.481 "name": "BaseBdev3", 00:08:08.481 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:08.481 "is_configured": true, 00:08:08.481 "data_offset": 2048, 00:08:08.481 "data_size": 63488 00:08:08.481 } 00:08:08.481 ] 00:08:08.481 }' 00:08:08.481 16:46:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.481 16:46:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.741 [2024-09-29 16:46:30.331824] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.741 "name": "Existed_Raid", 00:08:08.741 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:08.741 "strip_size_kb": 64, 00:08:08.741 "state": "configuring", 00:08:08.741 "raid_level": "raid0", 00:08:08.741 "superblock": true, 00:08:08.741 "num_base_bdevs": 3, 00:08:08.741 "num_base_bdevs_discovered": 2, 00:08:08.741 "num_base_bdevs_operational": 3, 00:08:08.741 "base_bdevs_list": [ 00:08:08.741 { 00:08:08.741 "name": null, 00:08:08.741 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:08.741 "is_configured": false, 00:08:08.741 "data_offset": 0, 00:08:08.741 "data_size": 63488 00:08:08.741 }, 00:08:08.741 { 00:08:08.741 "name": "BaseBdev2", 00:08:08.741 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:08.741 "is_configured": true, 00:08:08.741 "data_offset": 2048, 00:08:08.741 "data_size": 63488 00:08:08.741 }, 00:08:08.741 { 00:08:08.741 "name": "BaseBdev3", 00:08:08.741 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:08.741 "is_configured": true, 00:08:08.741 "data_offset": 2048, 00:08:08.741 "data_size": 63488 00:08:08.741 } 00:08:08.741 ] 00:08:08.741 }' 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.741 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 4182a219-21ea-41fa-a691-9c5c7624689c 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.312 [2024-09-29 16:46:30.865849] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:09.312 [2024-09-29 16:46:30.866012] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:09.312 [2024-09-29 16:46:30.866028] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:09.312 NewBaseBdev 00:08:09.312 [2024-09-29 16:46:30.866284] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:09.312 [2024-09-29 16:46:30.866401] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:09.312 [2024-09-29 16:46:30.866411] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:09.312 [2024-09-29 16:46:30.866518] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.312 [ 00:08:09.312 { 00:08:09.312 "name": "NewBaseBdev", 00:08:09.312 "aliases": [ 00:08:09.312 "4182a219-21ea-41fa-a691-9c5c7624689c" 00:08:09.312 ], 00:08:09.312 "product_name": "Malloc disk", 00:08:09.312 "block_size": 512, 00:08:09.312 "num_blocks": 65536, 00:08:09.312 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:09.312 "assigned_rate_limits": { 00:08:09.312 "rw_ios_per_sec": 0, 00:08:09.312 "rw_mbytes_per_sec": 0, 00:08:09.312 "r_mbytes_per_sec": 0, 00:08:09.312 "w_mbytes_per_sec": 0 00:08:09.312 }, 00:08:09.312 "claimed": true, 00:08:09.312 "claim_type": "exclusive_write", 00:08:09.312 "zoned": false, 00:08:09.312 "supported_io_types": { 00:08:09.312 "read": true, 00:08:09.312 "write": true, 00:08:09.312 "unmap": true, 00:08:09.312 "flush": true, 00:08:09.312 "reset": true, 00:08:09.312 "nvme_admin": false, 00:08:09.312 "nvme_io": false, 00:08:09.312 "nvme_io_md": false, 00:08:09.312 "write_zeroes": true, 00:08:09.312 "zcopy": true, 00:08:09.312 "get_zone_info": false, 00:08:09.312 "zone_management": false, 00:08:09.312 "zone_append": false, 00:08:09.312 "compare": false, 00:08:09.312 "compare_and_write": false, 00:08:09.312 "abort": true, 00:08:09.312 "seek_hole": false, 00:08:09.312 "seek_data": false, 00:08:09.312 "copy": true, 00:08:09.312 "nvme_iov_md": false 00:08:09.312 }, 00:08:09.312 "memory_domains": [ 00:08:09.312 { 00:08:09.312 "dma_device_id": "system", 00:08:09.312 "dma_device_type": 1 00:08:09.312 }, 00:08:09.312 { 00:08:09.312 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:09.312 "dma_device_type": 2 00:08:09.312 } 00:08:09.312 ], 00:08:09.312 "driver_specific": {} 00:08:09.312 } 00:08:09.312 ] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:09.312 "name": "Existed_Raid", 00:08:09.312 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:09.312 "strip_size_kb": 64, 00:08:09.312 "state": "online", 00:08:09.312 "raid_level": "raid0", 00:08:09.312 "superblock": true, 00:08:09.312 "num_base_bdevs": 3, 00:08:09.312 "num_base_bdevs_discovered": 3, 00:08:09.312 "num_base_bdevs_operational": 3, 00:08:09.312 "base_bdevs_list": [ 00:08:09.312 { 00:08:09.312 "name": "NewBaseBdev", 00:08:09.312 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:09.312 "is_configured": true, 00:08:09.312 "data_offset": 2048, 00:08:09.312 "data_size": 63488 00:08:09.312 }, 00:08:09.312 { 00:08:09.312 "name": "BaseBdev2", 00:08:09.312 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:09.312 "is_configured": true, 00:08:09.312 "data_offset": 2048, 00:08:09.312 "data_size": 63488 00:08:09.312 }, 00:08:09.312 { 00:08:09.312 "name": "BaseBdev3", 00:08:09.312 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:09.312 "is_configured": true, 00:08:09.312 "data_offset": 2048, 00:08:09.312 "data_size": 63488 00:08:09.312 } 00:08:09.312 ] 00:08:09.312 }' 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:09.312 16:46:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:09.882 [2024-09-29 16:46:31.357307] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.882 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:09.882 "name": "Existed_Raid", 00:08:09.882 "aliases": [ 00:08:09.882 "10c13813-d241-4ea5-a2ec-1ff033038a73" 00:08:09.882 ], 00:08:09.882 "product_name": "Raid Volume", 00:08:09.882 "block_size": 512, 00:08:09.882 "num_blocks": 190464, 00:08:09.882 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:09.882 "assigned_rate_limits": { 00:08:09.882 "rw_ios_per_sec": 0, 00:08:09.882 "rw_mbytes_per_sec": 0, 00:08:09.882 "r_mbytes_per_sec": 0, 00:08:09.882 "w_mbytes_per_sec": 0 00:08:09.882 }, 00:08:09.882 "claimed": false, 00:08:09.882 "zoned": false, 00:08:09.882 "supported_io_types": { 00:08:09.882 "read": true, 00:08:09.882 "write": true, 00:08:09.882 "unmap": true, 00:08:09.882 "flush": true, 00:08:09.882 "reset": true, 00:08:09.882 "nvme_admin": false, 00:08:09.882 "nvme_io": false, 00:08:09.882 "nvme_io_md": false, 00:08:09.882 "write_zeroes": true, 00:08:09.882 "zcopy": false, 00:08:09.882 "get_zone_info": false, 00:08:09.882 "zone_management": false, 00:08:09.882 "zone_append": false, 00:08:09.882 "compare": false, 00:08:09.882 "compare_and_write": false, 00:08:09.882 "abort": false, 00:08:09.882 "seek_hole": false, 00:08:09.882 "seek_data": false, 00:08:09.882 "copy": false, 00:08:09.882 "nvme_iov_md": false 00:08:09.882 }, 00:08:09.882 "memory_domains": [ 00:08:09.882 { 00:08:09.882 "dma_device_id": "system", 00:08:09.882 "dma_device_type": 1 00:08:09.882 }, 00:08:09.882 { 00:08:09.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:09.882 "dma_device_type": 2 00:08:09.882 }, 00:08:09.882 { 00:08:09.882 "dma_device_id": "system", 00:08:09.882 "dma_device_type": 1 00:08:09.882 }, 00:08:09.882 { 00:08:09.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:09.882 "dma_device_type": 2 00:08:09.882 }, 00:08:09.882 { 00:08:09.882 "dma_device_id": "system", 00:08:09.882 "dma_device_type": 1 00:08:09.882 }, 00:08:09.882 { 00:08:09.882 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:09.882 "dma_device_type": 2 00:08:09.882 } 00:08:09.882 ], 00:08:09.882 "driver_specific": { 00:08:09.882 "raid": { 00:08:09.882 "uuid": "10c13813-d241-4ea5-a2ec-1ff033038a73", 00:08:09.882 "strip_size_kb": 64, 00:08:09.882 "state": "online", 00:08:09.882 "raid_level": "raid0", 00:08:09.882 "superblock": true, 00:08:09.883 "num_base_bdevs": 3, 00:08:09.883 "num_base_bdevs_discovered": 3, 00:08:09.883 "num_base_bdevs_operational": 3, 00:08:09.883 "base_bdevs_list": [ 00:08:09.883 { 00:08:09.883 "name": "NewBaseBdev", 00:08:09.883 "uuid": "4182a219-21ea-41fa-a691-9c5c7624689c", 00:08:09.883 "is_configured": true, 00:08:09.883 "data_offset": 2048, 00:08:09.883 "data_size": 63488 00:08:09.883 }, 00:08:09.883 { 00:08:09.883 "name": "BaseBdev2", 00:08:09.883 "uuid": "5bf6560d-5d66-4389-9ec7-aed7bc0255b9", 00:08:09.883 "is_configured": true, 00:08:09.883 "data_offset": 2048, 00:08:09.883 "data_size": 63488 00:08:09.883 }, 00:08:09.883 { 00:08:09.883 "name": "BaseBdev3", 00:08:09.883 "uuid": "3b277c29-3b3e-4824-9c8e-fbf1c6b8dfb1", 00:08:09.883 "is_configured": true, 00:08:09.883 "data_offset": 2048, 00:08:09.883 "data_size": 63488 00:08:09.883 } 00:08:09.883 ] 00:08:09.883 } 00:08:09.883 } 00:08:09.883 }' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:09.883 BaseBdev2 00:08:09.883 BaseBdev3' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:09.883 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.143 [2024-09-29 16:46:31.636542] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:10.143 [2024-09-29 16:46:31.636566] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:10.143 [2024-09-29 16:46:31.636627] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:10.143 [2024-09-29 16:46:31.636674] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:10.143 [2024-09-29 16:46:31.636685] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75338 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 75338 ']' 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 75338 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75338 00:08:10.143 killing process with pid 75338 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75338' 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 75338 00:08:10.143 [2024-09-29 16:46:31.685630] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:10.143 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 75338 00:08:10.143 [2024-09-29 16:46:31.716394] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:10.403 16:46:31 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:10.403 00:08:10.403 real 0m8.891s 00:08:10.403 user 0m15.274s 00:08:10.403 sys 0m1.699s 00:08:10.403 ************************************ 00:08:10.403 END TEST raid_state_function_test_sb 00:08:10.403 ************************************ 00:08:10.403 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:10.403 16:46:31 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.403 16:46:32 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:08:10.403 16:46:32 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:10.403 16:46:32 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:10.403 16:46:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:10.403 ************************************ 00:08:10.403 START TEST raid_superblock_test 00:08:10.403 ************************************ 00:08:10.403 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 3 00:08:10.403 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:10.403 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:10.403 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:10.403 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=75947 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 75947 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 75947 ']' 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:10.404 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:10.404 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:10.664 [2024-09-29 16:46:32.106456] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:10.664 [2024-09-29 16:46:32.106672] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid75947 ] 00:08:10.664 [2024-09-29 16:46:32.250472] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:10.664 [2024-09-29 16:46:32.295799] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:10.923 [2024-09-29 16:46:32.337757] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:10.923 [2024-09-29 16:46:32.337795] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.493 malloc1 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.493 [2024-09-29 16:46:32.939913] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:11.493 [2024-09-29 16:46:32.940034] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:11.493 [2024-09-29 16:46:32.940068] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:11.493 [2024-09-29 16:46:32.940083] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:11.493 [2024-09-29 16:46:32.942158] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:11.493 [2024-09-29 16:46:32.942197] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:11.493 pt1 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.493 malloc2 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.493 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.493 [2024-09-29 16:46:32.978502] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:11.493 [2024-09-29 16:46:32.978608] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:11.494 [2024-09-29 16:46:32.978648] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:11.494 [2024-09-29 16:46:32.978685] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:11.494 [2024-09-29 16:46:32.981151] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:11.494 [2024-09-29 16:46:32.981235] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:11.494 pt2 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.494 16:46:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.494 malloc3 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.494 [2024-09-29 16:46:33.010824] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:11.494 [2024-09-29 16:46:33.010925] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:11.494 [2024-09-29 16:46:33.010959] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:11.494 [2024-09-29 16:46:33.010988] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:11.494 [2024-09-29 16:46:33.013017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:11.494 [2024-09-29 16:46:33.013090] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:11.494 pt3 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.494 [2024-09-29 16:46:33.022895] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:11.494 [2024-09-29 16:46:33.024713] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:11.494 [2024-09-29 16:46:33.024841] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:11.494 [2024-09-29 16:46:33.025016] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:11.494 [2024-09-29 16:46:33.025071] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:11.494 [2024-09-29 16:46:33.025353] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:11.494 [2024-09-29 16:46:33.025542] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:11.494 [2024-09-29 16:46:33.025590] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:11.494 [2024-09-29 16:46:33.025782] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.494 "name": "raid_bdev1", 00:08:11.494 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:11.494 "strip_size_kb": 64, 00:08:11.494 "state": "online", 00:08:11.494 "raid_level": "raid0", 00:08:11.494 "superblock": true, 00:08:11.494 "num_base_bdevs": 3, 00:08:11.494 "num_base_bdevs_discovered": 3, 00:08:11.494 "num_base_bdevs_operational": 3, 00:08:11.494 "base_bdevs_list": [ 00:08:11.494 { 00:08:11.494 "name": "pt1", 00:08:11.494 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:11.494 "is_configured": true, 00:08:11.494 "data_offset": 2048, 00:08:11.494 "data_size": 63488 00:08:11.494 }, 00:08:11.494 { 00:08:11.494 "name": "pt2", 00:08:11.494 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:11.494 "is_configured": true, 00:08:11.494 "data_offset": 2048, 00:08:11.494 "data_size": 63488 00:08:11.494 }, 00:08:11.494 { 00:08:11.494 "name": "pt3", 00:08:11.494 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:11.494 "is_configured": true, 00:08:11.494 "data_offset": 2048, 00:08:11.494 "data_size": 63488 00:08:11.494 } 00:08:11.494 ] 00:08:11.494 }' 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.494 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.064 [2024-09-29 16:46:33.494346] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:12.064 "name": "raid_bdev1", 00:08:12.064 "aliases": [ 00:08:12.064 "e1e47652-06a6-4c69-9c71-99c53a5c0d80" 00:08:12.064 ], 00:08:12.064 "product_name": "Raid Volume", 00:08:12.064 "block_size": 512, 00:08:12.064 "num_blocks": 190464, 00:08:12.064 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:12.064 "assigned_rate_limits": { 00:08:12.064 "rw_ios_per_sec": 0, 00:08:12.064 "rw_mbytes_per_sec": 0, 00:08:12.064 "r_mbytes_per_sec": 0, 00:08:12.064 "w_mbytes_per_sec": 0 00:08:12.064 }, 00:08:12.064 "claimed": false, 00:08:12.064 "zoned": false, 00:08:12.064 "supported_io_types": { 00:08:12.064 "read": true, 00:08:12.064 "write": true, 00:08:12.064 "unmap": true, 00:08:12.064 "flush": true, 00:08:12.064 "reset": true, 00:08:12.064 "nvme_admin": false, 00:08:12.064 "nvme_io": false, 00:08:12.064 "nvme_io_md": false, 00:08:12.064 "write_zeroes": true, 00:08:12.064 "zcopy": false, 00:08:12.064 "get_zone_info": false, 00:08:12.064 "zone_management": false, 00:08:12.064 "zone_append": false, 00:08:12.064 "compare": false, 00:08:12.064 "compare_and_write": false, 00:08:12.064 "abort": false, 00:08:12.064 "seek_hole": false, 00:08:12.064 "seek_data": false, 00:08:12.064 "copy": false, 00:08:12.064 "nvme_iov_md": false 00:08:12.064 }, 00:08:12.064 "memory_domains": [ 00:08:12.064 { 00:08:12.064 "dma_device_id": "system", 00:08:12.064 "dma_device_type": 1 00:08:12.064 }, 00:08:12.064 { 00:08:12.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.064 "dma_device_type": 2 00:08:12.064 }, 00:08:12.064 { 00:08:12.064 "dma_device_id": "system", 00:08:12.064 "dma_device_type": 1 00:08:12.064 }, 00:08:12.064 { 00:08:12.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.064 "dma_device_type": 2 00:08:12.064 }, 00:08:12.064 { 00:08:12.064 "dma_device_id": "system", 00:08:12.064 "dma_device_type": 1 00:08:12.064 }, 00:08:12.064 { 00:08:12.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.064 "dma_device_type": 2 00:08:12.064 } 00:08:12.064 ], 00:08:12.064 "driver_specific": { 00:08:12.064 "raid": { 00:08:12.064 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:12.064 "strip_size_kb": 64, 00:08:12.064 "state": "online", 00:08:12.064 "raid_level": "raid0", 00:08:12.064 "superblock": true, 00:08:12.064 "num_base_bdevs": 3, 00:08:12.064 "num_base_bdevs_discovered": 3, 00:08:12.064 "num_base_bdevs_operational": 3, 00:08:12.064 "base_bdevs_list": [ 00:08:12.064 { 00:08:12.064 "name": "pt1", 00:08:12.064 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:12.064 "is_configured": true, 00:08:12.064 "data_offset": 2048, 00:08:12.064 "data_size": 63488 00:08:12.064 }, 00:08:12.064 { 00:08:12.064 "name": "pt2", 00:08:12.064 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:12.064 "is_configured": true, 00:08:12.064 "data_offset": 2048, 00:08:12.064 "data_size": 63488 00:08:12.064 }, 00:08:12.064 { 00:08:12.064 "name": "pt3", 00:08:12.064 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:12.064 "is_configured": true, 00:08:12.064 "data_offset": 2048, 00:08:12.064 "data_size": 63488 00:08:12.064 } 00:08:12.064 ] 00:08:12.064 } 00:08:12.064 } 00:08:12.064 }' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:12.064 pt2 00:08:12.064 pt3' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.064 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 [2024-09-29 16:46:33.761873] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=e1e47652-06a6-4c69-9c71-99c53a5c0d80 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z e1e47652-06a6-4c69-9c71-99c53a5c0d80 ']' 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 [2024-09-29 16:46:33.793565] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:12.324 [2024-09-29 16:46:33.793632] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:12.324 [2024-09-29 16:46:33.793744] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:12.324 [2024-09-29 16:46:33.793818] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:12.324 [2024-09-29 16:46:33.793832] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.324 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.324 [2024-09-29 16:46:33.949332] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:12.324 [2024-09-29 16:46:33.951237] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:12.324 [2024-09-29 16:46:33.951282] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:12.325 [2024-09-29 16:46:33.951340] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:12.325 [2024-09-29 16:46:33.951383] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:12.325 [2024-09-29 16:46:33.951402] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:12.325 [2024-09-29 16:46:33.951415] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:12.325 [2024-09-29 16:46:33.951425] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:12.325 request: 00:08:12.325 { 00:08:12.325 "name": "raid_bdev1", 00:08:12.325 "raid_level": "raid0", 00:08:12.325 "base_bdevs": [ 00:08:12.325 "malloc1", 00:08:12.325 "malloc2", 00:08:12.325 "malloc3" 00:08:12.325 ], 00:08:12.325 "strip_size_kb": 64, 00:08:12.325 "superblock": false, 00:08:12.325 "method": "bdev_raid_create", 00:08:12.325 "req_id": 1 00:08:12.325 } 00:08:12.325 Got JSON-RPC error response 00:08:12.325 response: 00:08:12.325 { 00:08:12.325 "code": -17, 00:08:12.325 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:12.325 } 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.325 16:46:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.584 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.585 [2024-09-29 16:46:34.013176] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:12.585 [2024-09-29 16:46:34.013273] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.585 [2024-09-29 16:46:34.013305] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:12.585 [2024-09-29 16:46:34.013339] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.585 [2024-09-29 16:46:34.015484] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.585 [2024-09-29 16:46:34.015560] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:12.585 [2024-09-29 16:46:34.015656] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:12.585 [2024-09-29 16:46:34.015736] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:12.585 pt1 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.585 "name": "raid_bdev1", 00:08:12.585 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:12.585 "strip_size_kb": 64, 00:08:12.585 "state": "configuring", 00:08:12.585 "raid_level": "raid0", 00:08:12.585 "superblock": true, 00:08:12.585 "num_base_bdevs": 3, 00:08:12.585 "num_base_bdevs_discovered": 1, 00:08:12.585 "num_base_bdevs_operational": 3, 00:08:12.585 "base_bdevs_list": [ 00:08:12.585 { 00:08:12.585 "name": "pt1", 00:08:12.585 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:12.585 "is_configured": true, 00:08:12.585 "data_offset": 2048, 00:08:12.585 "data_size": 63488 00:08:12.585 }, 00:08:12.585 { 00:08:12.585 "name": null, 00:08:12.585 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:12.585 "is_configured": false, 00:08:12.585 "data_offset": 2048, 00:08:12.585 "data_size": 63488 00:08:12.585 }, 00:08:12.585 { 00:08:12.585 "name": null, 00:08:12.585 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:12.585 "is_configured": false, 00:08:12.585 "data_offset": 2048, 00:08:12.585 "data_size": 63488 00:08:12.585 } 00:08:12.585 ] 00:08:12.585 }' 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.585 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.845 [2024-09-29 16:46:34.428474] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:12.845 [2024-09-29 16:46:34.428537] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:12.845 [2024-09-29 16:46:34.428556] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:12.845 [2024-09-29 16:46:34.428569] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:12.845 [2024-09-29 16:46:34.428915] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:12.845 [2024-09-29 16:46:34.428933] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:12.845 [2024-09-29 16:46:34.428993] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:12.845 [2024-09-29 16:46:34.429016] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:12.845 pt2 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.845 [2024-09-29 16:46:34.440465] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.845 "name": "raid_bdev1", 00:08:12.845 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:12.845 "strip_size_kb": 64, 00:08:12.845 "state": "configuring", 00:08:12.845 "raid_level": "raid0", 00:08:12.845 "superblock": true, 00:08:12.845 "num_base_bdevs": 3, 00:08:12.845 "num_base_bdevs_discovered": 1, 00:08:12.845 "num_base_bdevs_operational": 3, 00:08:12.845 "base_bdevs_list": [ 00:08:12.845 { 00:08:12.845 "name": "pt1", 00:08:12.845 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:12.845 "is_configured": true, 00:08:12.845 "data_offset": 2048, 00:08:12.845 "data_size": 63488 00:08:12.845 }, 00:08:12.845 { 00:08:12.845 "name": null, 00:08:12.845 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:12.845 "is_configured": false, 00:08:12.845 "data_offset": 0, 00:08:12.845 "data_size": 63488 00:08:12.845 }, 00:08:12.845 { 00:08:12.845 "name": null, 00:08:12.845 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:12.845 "is_configured": false, 00:08:12.845 "data_offset": 2048, 00:08:12.845 "data_size": 63488 00:08:12.845 } 00:08:12.845 ] 00:08:12.845 }' 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.845 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.415 [2024-09-29 16:46:34.803828] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:13.415 [2024-09-29 16:46:34.803917] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.415 [2024-09-29 16:46:34.803952] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:13.415 [2024-09-29 16:46:34.803979] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.415 [2024-09-29 16:46:34.804353] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.415 [2024-09-29 16:46:34.804407] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:13.415 [2024-09-29 16:46:34.804504] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:13.415 [2024-09-29 16:46:34.804550] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:13.415 pt2 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.415 [2024-09-29 16:46:34.815827] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:13.415 [2024-09-29 16:46:34.815902] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:13.415 [2024-09-29 16:46:34.815935] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:13.415 [2024-09-29 16:46:34.815960] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:13.415 [2024-09-29 16:46:34.816285] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:13.415 [2024-09-29 16:46:34.816338] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:13.415 [2024-09-29 16:46:34.816430] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:13.415 [2024-09-29 16:46:34.816485] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:13.415 [2024-09-29 16:46:34.816620] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:13.415 [2024-09-29 16:46:34.816657] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:13.415 [2024-09-29 16:46:34.816920] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:13.415 [2024-09-29 16:46:34.817066] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:13.415 [2024-09-29 16:46:34.817106] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:13.415 [2024-09-29 16:46:34.817250] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:13.415 pt3 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:13.415 "name": "raid_bdev1", 00:08:13.415 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:13.415 "strip_size_kb": 64, 00:08:13.415 "state": "online", 00:08:13.415 "raid_level": "raid0", 00:08:13.415 "superblock": true, 00:08:13.415 "num_base_bdevs": 3, 00:08:13.415 "num_base_bdevs_discovered": 3, 00:08:13.415 "num_base_bdevs_operational": 3, 00:08:13.415 "base_bdevs_list": [ 00:08:13.415 { 00:08:13.415 "name": "pt1", 00:08:13.415 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:13.415 "is_configured": true, 00:08:13.415 "data_offset": 2048, 00:08:13.415 "data_size": 63488 00:08:13.415 }, 00:08:13.415 { 00:08:13.415 "name": "pt2", 00:08:13.415 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:13.415 "is_configured": true, 00:08:13.415 "data_offset": 2048, 00:08:13.415 "data_size": 63488 00:08:13.415 }, 00:08:13.415 { 00:08:13.415 "name": "pt3", 00:08:13.415 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:13.415 "is_configured": true, 00:08:13.415 "data_offset": 2048, 00:08:13.415 "data_size": 63488 00:08:13.415 } 00:08:13.415 ] 00:08:13.415 }' 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:13.415 16:46:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.675 [2024-09-29 16:46:35.243415] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.675 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:13.675 "name": "raid_bdev1", 00:08:13.675 "aliases": [ 00:08:13.675 "e1e47652-06a6-4c69-9c71-99c53a5c0d80" 00:08:13.675 ], 00:08:13.675 "product_name": "Raid Volume", 00:08:13.675 "block_size": 512, 00:08:13.675 "num_blocks": 190464, 00:08:13.675 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:13.675 "assigned_rate_limits": { 00:08:13.675 "rw_ios_per_sec": 0, 00:08:13.675 "rw_mbytes_per_sec": 0, 00:08:13.675 "r_mbytes_per_sec": 0, 00:08:13.675 "w_mbytes_per_sec": 0 00:08:13.675 }, 00:08:13.675 "claimed": false, 00:08:13.675 "zoned": false, 00:08:13.675 "supported_io_types": { 00:08:13.675 "read": true, 00:08:13.675 "write": true, 00:08:13.675 "unmap": true, 00:08:13.675 "flush": true, 00:08:13.675 "reset": true, 00:08:13.675 "nvme_admin": false, 00:08:13.675 "nvme_io": false, 00:08:13.676 "nvme_io_md": false, 00:08:13.676 "write_zeroes": true, 00:08:13.676 "zcopy": false, 00:08:13.676 "get_zone_info": false, 00:08:13.676 "zone_management": false, 00:08:13.676 "zone_append": false, 00:08:13.676 "compare": false, 00:08:13.676 "compare_and_write": false, 00:08:13.676 "abort": false, 00:08:13.676 "seek_hole": false, 00:08:13.676 "seek_data": false, 00:08:13.676 "copy": false, 00:08:13.676 "nvme_iov_md": false 00:08:13.676 }, 00:08:13.676 "memory_domains": [ 00:08:13.676 { 00:08:13.676 "dma_device_id": "system", 00:08:13.676 "dma_device_type": 1 00:08:13.676 }, 00:08:13.676 { 00:08:13.676 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.676 "dma_device_type": 2 00:08:13.676 }, 00:08:13.676 { 00:08:13.676 "dma_device_id": "system", 00:08:13.676 "dma_device_type": 1 00:08:13.676 }, 00:08:13.676 { 00:08:13.676 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.676 "dma_device_type": 2 00:08:13.676 }, 00:08:13.676 { 00:08:13.676 "dma_device_id": "system", 00:08:13.676 "dma_device_type": 1 00:08:13.676 }, 00:08:13.676 { 00:08:13.676 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:13.676 "dma_device_type": 2 00:08:13.676 } 00:08:13.676 ], 00:08:13.676 "driver_specific": { 00:08:13.676 "raid": { 00:08:13.676 "uuid": "e1e47652-06a6-4c69-9c71-99c53a5c0d80", 00:08:13.676 "strip_size_kb": 64, 00:08:13.676 "state": "online", 00:08:13.676 "raid_level": "raid0", 00:08:13.676 "superblock": true, 00:08:13.676 "num_base_bdevs": 3, 00:08:13.676 "num_base_bdevs_discovered": 3, 00:08:13.676 "num_base_bdevs_operational": 3, 00:08:13.676 "base_bdevs_list": [ 00:08:13.676 { 00:08:13.676 "name": "pt1", 00:08:13.676 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:13.676 "is_configured": true, 00:08:13.676 "data_offset": 2048, 00:08:13.676 "data_size": 63488 00:08:13.676 }, 00:08:13.676 { 00:08:13.676 "name": "pt2", 00:08:13.676 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:13.676 "is_configured": true, 00:08:13.676 "data_offset": 2048, 00:08:13.676 "data_size": 63488 00:08:13.676 }, 00:08:13.676 { 00:08:13.676 "name": "pt3", 00:08:13.676 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:13.676 "is_configured": true, 00:08:13.676 "data_offset": 2048, 00:08:13.676 "data_size": 63488 00:08:13.676 } 00:08:13.676 ] 00:08:13.676 } 00:08:13.676 } 00:08:13.676 }' 00:08:13.676 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:13.676 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:13.676 pt2 00:08:13.676 pt3' 00:08:13.676 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.936 [2024-09-29 16:46:35.514886] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' e1e47652-06a6-4c69-9c71-99c53a5c0d80 '!=' e1e47652-06a6-4c69-9c71-99c53a5c0d80 ']' 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:13.936 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 75947 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 75947 ']' 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 75947 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75947 00:08:13.937 killing process with pid 75947 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75947' 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 75947 00:08:13.937 [2024-09-29 16:46:35.597694] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:13.937 [2024-09-29 16:46:35.597780] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:13.937 [2024-09-29 16:46:35.597841] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:13.937 [2024-09-29 16:46:35.597851] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:13.937 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 75947 00:08:14.196 [2024-09-29 16:46:35.630268] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:14.457 16:46:35 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:14.457 00:08:14.457 real 0m3.844s 00:08:14.457 user 0m6.038s 00:08:14.457 sys 0m0.833s 00:08:14.457 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:14.457 16:46:35 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.457 ************************************ 00:08:14.457 END TEST raid_superblock_test 00:08:14.457 ************************************ 00:08:14.457 16:46:35 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:08:14.457 16:46:35 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:14.457 16:46:35 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:14.457 16:46:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:14.457 ************************************ 00:08:14.457 START TEST raid_read_error_test 00:08:14.457 ************************************ 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 read 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.QTIkqiGa36 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76184 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76184 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 76184 ']' 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:14.457 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:14.457 16:46:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.457 [2024-09-29 16:46:36.039588] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:14.457 [2024-09-29 16:46:36.039703] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76184 ] 00:08:14.717 [2024-09-29 16:46:36.181853] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:14.717 [2024-09-29 16:46:36.227412] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:14.717 [2024-09-29 16:46:36.269210] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:14.717 [2024-09-29 16:46:36.269256] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.287 BaseBdev1_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.287 true 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.287 [2024-09-29 16:46:36.878977] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:15.287 [2024-09-29 16:46:36.879033] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.287 [2024-09-29 16:46:36.879054] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:15.287 [2024-09-29 16:46:36.879062] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.287 [2024-09-29 16:46:36.881158] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.287 [2024-09-29 16:46:36.881194] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:15.287 BaseBdev1 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.287 BaseBdev2_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.287 true 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.287 [2024-09-29 16:46:36.931844] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:15.287 [2024-09-29 16:46:36.931928] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.287 [2024-09-29 16:46:36.931963] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:15.287 [2024-09-29 16:46:36.931978] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.287 [2024-09-29 16:46:36.934796] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.287 [2024-09-29 16:46:36.934837] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:15.287 BaseBdev2 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.287 BaseBdev3_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.287 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.548 true 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.548 [2024-09-29 16:46:36.972252] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:15.548 [2024-09-29 16:46:36.972351] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.548 [2024-09-29 16:46:36.972373] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:15.548 [2024-09-29 16:46:36.972382] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.548 [2024-09-29 16:46:36.974740] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.548 [2024-09-29 16:46:36.974785] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:15.548 BaseBdev3 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.548 [2024-09-29 16:46:36.984325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:15.548 [2024-09-29 16:46:36.986110] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:15.548 [2024-09-29 16:46:36.986181] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:15.548 [2024-09-29 16:46:36.986342] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:15.548 [2024-09-29 16:46:36.986362] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:15.548 [2024-09-29 16:46:36.986607] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:15.548 [2024-09-29 16:46:36.986745] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:15.548 [2024-09-29 16:46:36.986757] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:15.548 [2024-09-29 16:46:36.986901] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.548 16:46:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.548 16:46:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.548 16:46:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.548 "name": "raid_bdev1", 00:08:15.548 "uuid": "a10afb1e-7cac-43ed-bdca-fd76d5d8ae82", 00:08:15.548 "strip_size_kb": 64, 00:08:15.548 "state": "online", 00:08:15.548 "raid_level": "raid0", 00:08:15.548 "superblock": true, 00:08:15.548 "num_base_bdevs": 3, 00:08:15.548 "num_base_bdevs_discovered": 3, 00:08:15.548 "num_base_bdevs_operational": 3, 00:08:15.548 "base_bdevs_list": [ 00:08:15.548 { 00:08:15.548 "name": "BaseBdev1", 00:08:15.548 "uuid": "2265ffd5-d608-5b8f-a5f9-6c5a5632444c", 00:08:15.548 "is_configured": true, 00:08:15.548 "data_offset": 2048, 00:08:15.548 "data_size": 63488 00:08:15.548 }, 00:08:15.548 { 00:08:15.548 "name": "BaseBdev2", 00:08:15.548 "uuid": "ea7bc5be-dcad-5b6a-a372-f8bd6af19a00", 00:08:15.548 "is_configured": true, 00:08:15.548 "data_offset": 2048, 00:08:15.548 "data_size": 63488 00:08:15.548 }, 00:08:15.548 { 00:08:15.548 "name": "BaseBdev3", 00:08:15.548 "uuid": "a92d18c5-2d79-5270-8862-c77fbae8c2f2", 00:08:15.548 "is_configured": true, 00:08:15.548 "data_offset": 2048, 00:08:15.548 "data_size": 63488 00:08:15.548 } 00:08:15.548 ] 00:08:15.548 }' 00:08:15.548 16:46:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.548 16:46:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.808 16:46:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:15.809 16:46:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:16.068 [2024-09-29 16:46:37.515825] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.013 "name": "raid_bdev1", 00:08:17.013 "uuid": "a10afb1e-7cac-43ed-bdca-fd76d5d8ae82", 00:08:17.013 "strip_size_kb": 64, 00:08:17.013 "state": "online", 00:08:17.013 "raid_level": "raid0", 00:08:17.013 "superblock": true, 00:08:17.013 "num_base_bdevs": 3, 00:08:17.013 "num_base_bdevs_discovered": 3, 00:08:17.013 "num_base_bdevs_operational": 3, 00:08:17.013 "base_bdevs_list": [ 00:08:17.013 { 00:08:17.013 "name": "BaseBdev1", 00:08:17.013 "uuid": "2265ffd5-d608-5b8f-a5f9-6c5a5632444c", 00:08:17.013 "is_configured": true, 00:08:17.013 "data_offset": 2048, 00:08:17.013 "data_size": 63488 00:08:17.013 }, 00:08:17.013 { 00:08:17.013 "name": "BaseBdev2", 00:08:17.013 "uuid": "ea7bc5be-dcad-5b6a-a372-f8bd6af19a00", 00:08:17.013 "is_configured": true, 00:08:17.013 "data_offset": 2048, 00:08:17.013 "data_size": 63488 00:08:17.013 }, 00:08:17.013 { 00:08:17.013 "name": "BaseBdev3", 00:08:17.013 "uuid": "a92d18c5-2d79-5270-8862-c77fbae8c2f2", 00:08:17.013 "is_configured": true, 00:08:17.013 "data_offset": 2048, 00:08:17.013 "data_size": 63488 00:08:17.013 } 00:08:17.013 ] 00:08:17.013 }' 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.013 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.273 [2024-09-29 16:46:38.883379] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:17.273 [2024-09-29 16:46:38.883479] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:17.273 [2024-09-29 16:46:38.886021] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:17.273 [2024-09-29 16:46:38.886068] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:17.273 [2024-09-29 16:46:38.886101] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:17.273 [2024-09-29 16:46:38.886111] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:17.273 { 00:08:17.273 "results": [ 00:08:17.273 { 00:08:17.273 "job": "raid_bdev1", 00:08:17.273 "core_mask": "0x1", 00:08:17.273 "workload": "randrw", 00:08:17.273 "percentage": 50, 00:08:17.273 "status": "finished", 00:08:17.273 "queue_depth": 1, 00:08:17.273 "io_size": 131072, 00:08:17.273 "runtime": 1.368415, 00:08:17.273 "iops": 17333.192050657148, 00:08:17.273 "mibps": 2166.6490063321435, 00:08:17.273 "io_failed": 1, 00:08:17.273 "io_timeout": 0, 00:08:17.273 "avg_latency_us": 79.89023144841197, 00:08:17.273 "min_latency_us": 24.482096069868994, 00:08:17.273 "max_latency_us": 1352.216593886463 00:08:17.273 } 00:08:17.273 ], 00:08:17.273 "core_count": 1 00:08:17.273 } 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76184 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 76184 ']' 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 76184 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76184 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76184' 00:08:17.273 killing process with pid 76184 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 76184 00:08:17.273 [2024-09-29 16:46:38.924964] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:17.273 16:46:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 76184 00:08:17.532 [2024-09-29 16:46:38.950099] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.QTIkqiGa36 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:17.532 00:08:17.532 real 0m3.253s 00:08:17.532 user 0m4.101s 00:08:17.532 sys 0m0.513s 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:17.532 16:46:39 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.532 ************************************ 00:08:17.532 END TEST raid_read_error_test 00:08:17.532 ************************************ 00:08:17.791 16:46:39 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:08:17.791 16:46:39 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:17.791 16:46:39 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:17.791 16:46:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:17.791 ************************************ 00:08:17.791 START TEST raid_write_error_test 00:08:17.791 ************************************ 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 write 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.7poZyQBuAC 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76319 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76319 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 76319 ']' 00:08:17.791 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:17.791 16:46:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.791 [2024-09-29 16:46:39.360675] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:17.791 [2024-09-29 16:46:39.360899] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76319 ] 00:08:18.051 [2024-09-29 16:46:39.504857] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:18.051 [2024-09-29 16:46:39.550678] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.051 [2024-09-29 16:46:39.592712] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.051 [2024-09-29 16:46:39.592753] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.621 BaseBdev1_malloc 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.621 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.622 true 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.622 [2024-09-29 16:46:40.214574] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:18.622 [2024-09-29 16:46:40.214636] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:18.622 [2024-09-29 16:46:40.214659] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:18.622 [2024-09-29 16:46:40.214668] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:18.622 [2024-09-29 16:46:40.216779] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:18.622 [2024-09-29 16:46:40.216870] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:18.622 BaseBdev1 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.622 BaseBdev2_malloc 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.622 true 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.622 [2024-09-29 16:46:40.265321] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:18.622 [2024-09-29 16:46:40.265430] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:18.622 [2024-09-29 16:46:40.265454] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:18.622 [2024-09-29 16:46:40.265462] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:18.622 [2024-09-29 16:46:40.267527] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:18.622 [2024-09-29 16:46:40.267563] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:18.622 BaseBdev2 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.622 BaseBdev3_malloc 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.622 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.889 true 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.889 [2024-09-29 16:46:40.305911] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:18.889 [2024-09-29 16:46:40.305954] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:18.889 [2024-09-29 16:46:40.305972] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:18.889 [2024-09-29 16:46:40.305980] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:18.889 [2024-09-29 16:46:40.307988] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:18.889 [2024-09-29 16:46:40.308024] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:18.889 BaseBdev3 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.889 [2024-09-29 16:46:40.317974] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:18.889 [2024-09-29 16:46:40.319778] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:18.889 [2024-09-29 16:46:40.319853] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:18.889 [2024-09-29 16:46:40.320036] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:18.889 [2024-09-29 16:46:40.320058] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:18.889 [2024-09-29 16:46:40.320301] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:18.889 [2024-09-29 16:46:40.320446] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:18.889 [2024-09-29 16:46:40.320456] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:18.889 [2024-09-29 16:46:40.320585] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:18.889 "name": "raid_bdev1", 00:08:18.889 "uuid": "01549c68-9b05-4491-92d6-5fe936fb7ab8", 00:08:18.889 "strip_size_kb": 64, 00:08:18.889 "state": "online", 00:08:18.889 "raid_level": "raid0", 00:08:18.889 "superblock": true, 00:08:18.889 "num_base_bdevs": 3, 00:08:18.889 "num_base_bdevs_discovered": 3, 00:08:18.889 "num_base_bdevs_operational": 3, 00:08:18.889 "base_bdevs_list": [ 00:08:18.889 { 00:08:18.889 "name": "BaseBdev1", 00:08:18.889 "uuid": "0dda6c17-19aa-512d-9979-5639e0a58bd0", 00:08:18.889 "is_configured": true, 00:08:18.889 "data_offset": 2048, 00:08:18.889 "data_size": 63488 00:08:18.889 }, 00:08:18.889 { 00:08:18.889 "name": "BaseBdev2", 00:08:18.889 "uuid": "76733213-bb64-5e5f-9953-eef532b09f92", 00:08:18.889 "is_configured": true, 00:08:18.889 "data_offset": 2048, 00:08:18.889 "data_size": 63488 00:08:18.889 }, 00:08:18.889 { 00:08:18.889 "name": "BaseBdev3", 00:08:18.889 "uuid": "5dbe7944-6811-5c23-9d6a-843f391ce4fb", 00:08:18.889 "is_configured": true, 00:08:18.889 "data_offset": 2048, 00:08:18.889 "data_size": 63488 00:08:18.889 } 00:08:18.889 ] 00:08:18.889 }' 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:18.889 16:46:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.185 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:19.185 16:46:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:19.185 [2024-09-29 16:46:40.813451] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:20.122 "name": "raid_bdev1", 00:08:20.122 "uuid": "01549c68-9b05-4491-92d6-5fe936fb7ab8", 00:08:20.122 "strip_size_kb": 64, 00:08:20.122 "state": "online", 00:08:20.122 "raid_level": "raid0", 00:08:20.122 "superblock": true, 00:08:20.122 "num_base_bdevs": 3, 00:08:20.122 "num_base_bdevs_discovered": 3, 00:08:20.122 "num_base_bdevs_operational": 3, 00:08:20.122 "base_bdevs_list": [ 00:08:20.122 { 00:08:20.122 "name": "BaseBdev1", 00:08:20.122 "uuid": "0dda6c17-19aa-512d-9979-5639e0a58bd0", 00:08:20.122 "is_configured": true, 00:08:20.122 "data_offset": 2048, 00:08:20.122 "data_size": 63488 00:08:20.122 }, 00:08:20.122 { 00:08:20.122 "name": "BaseBdev2", 00:08:20.122 "uuid": "76733213-bb64-5e5f-9953-eef532b09f92", 00:08:20.122 "is_configured": true, 00:08:20.122 "data_offset": 2048, 00:08:20.122 "data_size": 63488 00:08:20.122 }, 00:08:20.122 { 00:08:20.122 "name": "BaseBdev3", 00:08:20.122 "uuid": "5dbe7944-6811-5c23-9d6a-843f391ce4fb", 00:08:20.122 "is_configured": true, 00:08:20.122 "data_offset": 2048, 00:08:20.122 "data_size": 63488 00:08:20.122 } 00:08:20.122 ] 00:08:20.122 }' 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:20.122 16:46:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.691 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:20.691 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.691 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.691 [2024-09-29 16:46:42.229676] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:20.692 [2024-09-29 16:46:42.229710] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:20.692 [2024-09-29 16:46:42.232177] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:20.692 [2024-09-29 16:46:42.232225] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:20.692 [2024-09-29 16:46:42.232259] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:20.692 [2024-09-29 16:46:42.232269] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:20.692 { 00:08:20.692 "results": [ 00:08:20.692 { 00:08:20.692 "job": "raid_bdev1", 00:08:20.692 "core_mask": "0x1", 00:08:20.692 "workload": "randrw", 00:08:20.692 "percentage": 50, 00:08:20.692 "status": "finished", 00:08:20.692 "queue_depth": 1, 00:08:20.692 "io_size": 131072, 00:08:20.692 "runtime": 1.417111, 00:08:20.692 "iops": 17271.75923410375, 00:08:20.692 "mibps": 2158.9699042629686, 00:08:20.692 "io_failed": 1, 00:08:20.692 "io_timeout": 0, 00:08:20.692 "avg_latency_us": 80.30056199269504, 00:08:20.692 "min_latency_us": 19.89868995633188, 00:08:20.692 "max_latency_us": 1359.3711790393013 00:08:20.692 } 00:08:20.692 ], 00:08:20.692 "core_count": 1 00:08:20.692 } 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76319 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 76319 ']' 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 76319 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76319 00:08:20.692 killing process with pid 76319 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76319' 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 76319 00:08:20.692 [2024-09-29 16:46:42.275969] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:20.692 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 76319 00:08:20.692 [2024-09-29 16:46:42.300615] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.7poZyQBuAC 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:08:20.951 00:08:20.951 real 0m3.278s 00:08:20.951 user 0m4.168s 00:08:20.951 sys 0m0.500s 00:08:20.951 ************************************ 00:08:20.951 END TEST raid_write_error_test 00:08:20.951 ************************************ 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:20.951 16:46:42 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:20.951 16:46:42 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:20.951 16:46:42 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:08:20.951 16:46:42 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:20.951 16:46:42 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:20.951 16:46:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:20.951 ************************************ 00:08:20.951 START TEST raid_state_function_test 00:08:20.951 ************************************ 00:08:20.951 16:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 false 00:08:20.951 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:20.951 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:20.952 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76446 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76446' 00:08:21.211 Process raid pid: 76446 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76446 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 76446 ']' 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:21.211 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:21.211 16:46:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.211 [2024-09-29 16:46:42.701470] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:21.211 [2024-09-29 16:46:42.701690] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:21.211 [2024-09-29 16:46:42.846800] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:21.470 [2024-09-29 16:46:42.893300] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:21.470 [2024-09-29 16:46:42.935155] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:21.470 [2024-09-29 16:46:42.935263] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.038 [2024-09-29 16:46:43.528356] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:22.038 [2024-09-29 16:46:43.528452] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:22.038 [2024-09-29 16:46:43.528497] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:22.038 [2024-09-29 16:46:43.528522] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:22.038 [2024-09-29 16:46:43.528540] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:22.038 [2024-09-29 16:46:43.528564] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.038 "name": "Existed_Raid", 00:08:22.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.038 "strip_size_kb": 64, 00:08:22.038 "state": "configuring", 00:08:22.038 "raid_level": "concat", 00:08:22.038 "superblock": false, 00:08:22.038 "num_base_bdevs": 3, 00:08:22.038 "num_base_bdevs_discovered": 0, 00:08:22.038 "num_base_bdevs_operational": 3, 00:08:22.038 "base_bdevs_list": [ 00:08:22.038 { 00:08:22.038 "name": "BaseBdev1", 00:08:22.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.038 "is_configured": false, 00:08:22.038 "data_offset": 0, 00:08:22.038 "data_size": 0 00:08:22.038 }, 00:08:22.038 { 00:08:22.038 "name": "BaseBdev2", 00:08:22.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.038 "is_configured": false, 00:08:22.038 "data_offset": 0, 00:08:22.038 "data_size": 0 00:08:22.038 }, 00:08:22.038 { 00:08:22.038 "name": "BaseBdev3", 00:08:22.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.038 "is_configured": false, 00:08:22.038 "data_offset": 0, 00:08:22.038 "data_size": 0 00:08:22.038 } 00:08:22.038 ] 00:08:22.038 }' 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.038 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.297 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:22.297 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.297 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.297 [2024-09-29 16:46:43.967497] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:22.297 [2024-09-29 16:46:43.967588] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:22.555 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.555 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:22.555 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.555 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.555 [2024-09-29 16:46:43.979495] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:22.555 [2024-09-29 16:46:43.979570] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:22.555 [2024-09-29 16:46:43.979597] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:22.555 [2024-09-29 16:46:43.979619] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:22.555 [2024-09-29 16:46:43.979636] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:22.555 [2024-09-29 16:46:43.979656] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:22.555 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.555 16:46:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:22.556 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.556 16:46:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.556 [2024-09-29 16:46:44.000408] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:22.556 BaseBdev1 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.556 [ 00:08:22.556 { 00:08:22.556 "name": "BaseBdev1", 00:08:22.556 "aliases": [ 00:08:22.556 "4bf8b6e0-624b-434c-a1bf-b08afd490a46" 00:08:22.556 ], 00:08:22.556 "product_name": "Malloc disk", 00:08:22.556 "block_size": 512, 00:08:22.556 "num_blocks": 65536, 00:08:22.556 "uuid": "4bf8b6e0-624b-434c-a1bf-b08afd490a46", 00:08:22.556 "assigned_rate_limits": { 00:08:22.556 "rw_ios_per_sec": 0, 00:08:22.556 "rw_mbytes_per_sec": 0, 00:08:22.556 "r_mbytes_per_sec": 0, 00:08:22.556 "w_mbytes_per_sec": 0 00:08:22.556 }, 00:08:22.556 "claimed": true, 00:08:22.556 "claim_type": "exclusive_write", 00:08:22.556 "zoned": false, 00:08:22.556 "supported_io_types": { 00:08:22.556 "read": true, 00:08:22.556 "write": true, 00:08:22.556 "unmap": true, 00:08:22.556 "flush": true, 00:08:22.556 "reset": true, 00:08:22.556 "nvme_admin": false, 00:08:22.556 "nvme_io": false, 00:08:22.556 "nvme_io_md": false, 00:08:22.556 "write_zeroes": true, 00:08:22.556 "zcopy": true, 00:08:22.556 "get_zone_info": false, 00:08:22.556 "zone_management": false, 00:08:22.556 "zone_append": false, 00:08:22.556 "compare": false, 00:08:22.556 "compare_and_write": false, 00:08:22.556 "abort": true, 00:08:22.556 "seek_hole": false, 00:08:22.556 "seek_data": false, 00:08:22.556 "copy": true, 00:08:22.556 "nvme_iov_md": false 00:08:22.556 }, 00:08:22.556 "memory_domains": [ 00:08:22.556 { 00:08:22.556 "dma_device_id": "system", 00:08:22.556 "dma_device_type": 1 00:08:22.556 }, 00:08:22.556 { 00:08:22.556 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:22.556 "dma_device_type": 2 00:08:22.556 } 00:08:22.556 ], 00:08:22.556 "driver_specific": {} 00:08:22.556 } 00:08:22.556 ] 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:22.556 "name": "Existed_Raid", 00:08:22.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.556 "strip_size_kb": 64, 00:08:22.556 "state": "configuring", 00:08:22.556 "raid_level": "concat", 00:08:22.556 "superblock": false, 00:08:22.556 "num_base_bdevs": 3, 00:08:22.556 "num_base_bdevs_discovered": 1, 00:08:22.556 "num_base_bdevs_operational": 3, 00:08:22.556 "base_bdevs_list": [ 00:08:22.556 { 00:08:22.556 "name": "BaseBdev1", 00:08:22.556 "uuid": "4bf8b6e0-624b-434c-a1bf-b08afd490a46", 00:08:22.556 "is_configured": true, 00:08:22.556 "data_offset": 0, 00:08:22.556 "data_size": 65536 00:08:22.556 }, 00:08:22.556 { 00:08:22.556 "name": "BaseBdev2", 00:08:22.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.556 "is_configured": false, 00:08:22.556 "data_offset": 0, 00:08:22.556 "data_size": 0 00:08:22.556 }, 00:08:22.556 { 00:08:22.556 "name": "BaseBdev3", 00:08:22.556 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:22.556 "is_configured": false, 00:08:22.556 "data_offset": 0, 00:08:22.556 "data_size": 0 00:08:22.556 } 00:08:22.556 ] 00:08:22.556 }' 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:22.556 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.815 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:22.815 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.815 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.815 [2024-09-29 16:46:44.463656] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:22.815 [2024-09-29 16:46:44.463758] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:22.815 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.815 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:22.815 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.815 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.815 [2024-09-29 16:46:44.475689] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:22.815 [2024-09-29 16:46:44.477619] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:22.815 [2024-09-29 16:46:44.477694] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:22.816 [2024-09-29 16:46:44.477729] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:22.816 [2024-09-29 16:46:44.477754] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:22.816 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.075 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.075 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.075 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.075 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.075 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.075 "name": "Existed_Raid", 00:08:23.075 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.075 "strip_size_kb": 64, 00:08:23.075 "state": "configuring", 00:08:23.075 "raid_level": "concat", 00:08:23.075 "superblock": false, 00:08:23.075 "num_base_bdevs": 3, 00:08:23.075 "num_base_bdevs_discovered": 1, 00:08:23.075 "num_base_bdevs_operational": 3, 00:08:23.075 "base_bdevs_list": [ 00:08:23.075 { 00:08:23.075 "name": "BaseBdev1", 00:08:23.075 "uuid": "4bf8b6e0-624b-434c-a1bf-b08afd490a46", 00:08:23.075 "is_configured": true, 00:08:23.075 "data_offset": 0, 00:08:23.075 "data_size": 65536 00:08:23.075 }, 00:08:23.075 { 00:08:23.075 "name": "BaseBdev2", 00:08:23.075 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.075 "is_configured": false, 00:08:23.075 "data_offset": 0, 00:08:23.075 "data_size": 0 00:08:23.075 }, 00:08:23.075 { 00:08:23.075 "name": "BaseBdev3", 00:08:23.075 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.075 "is_configured": false, 00:08:23.075 "data_offset": 0, 00:08:23.075 "data_size": 0 00:08:23.075 } 00:08:23.075 ] 00:08:23.075 }' 00:08:23.075 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.075 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.334 [2024-09-29 16:46:44.961399] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:23.334 BaseBdev2 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.334 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.334 [ 00:08:23.334 { 00:08:23.334 "name": "BaseBdev2", 00:08:23.334 "aliases": [ 00:08:23.334 "0e99e2e0-28ca-4064-a592-b4673c31ba4c" 00:08:23.334 ], 00:08:23.334 "product_name": "Malloc disk", 00:08:23.334 "block_size": 512, 00:08:23.334 "num_blocks": 65536, 00:08:23.334 "uuid": "0e99e2e0-28ca-4064-a592-b4673c31ba4c", 00:08:23.334 "assigned_rate_limits": { 00:08:23.334 "rw_ios_per_sec": 0, 00:08:23.334 "rw_mbytes_per_sec": 0, 00:08:23.334 "r_mbytes_per_sec": 0, 00:08:23.334 "w_mbytes_per_sec": 0 00:08:23.334 }, 00:08:23.334 "claimed": true, 00:08:23.334 "claim_type": "exclusive_write", 00:08:23.334 "zoned": false, 00:08:23.335 "supported_io_types": { 00:08:23.335 "read": true, 00:08:23.335 "write": true, 00:08:23.335 "unmap": true, 00:08:23.335 "flush": true, 00:08:23.335 "reset": true, 00:08:23.335 "nvme_admin": false, 00:08:23.335 "nvme_io": false, 00:08:23.335 "nvme_io_md": false, 00:08:23.335 "write_zeroes": true, 00:08:23.335 "zcopy": true, 00:08:23.335 "get_zone_info": false, 00:08:23.335 "zone_management": false, 00:08:23.335 "zone_append": false, 00:08:23.335 "compare": false, 00:08:23.335 "compare_and_write": false, 00:08:23.335 "abort": true, 00:08:23.335 "seek_hole": false, 00:08:23.335 "seek_data": false, 00:08:23.335 "copy": true, 00:08:23.335 "nvme_iov_md": false 00:08:23.335 }, 00:08:23.335 "memory_domains": [ 00:08:23.335 { 00:08:23.335 "dma_device_id": "system", 00:08:23.335 "dma_device_type": 1 00:08:23.335 }, 00:08:23.335 { 00:08:23.335 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.335 "dma_device_type": 2 00:08:23.335 } 00:08:23.335 ], 00:08:23.335 "driver_specific": {} 00:08:23.335 } 00:08:23.335 ] 00:08:23.335 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.335 16:46:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:23.335 16:46:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.335 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.594 "name": "Existed_Raid", 00:08:23.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.594 "strip_size_kb": 64, 00:08:23.594 "state": "configuring", 00:08:23.594 "raid_level": "concat", 00:08:23.594 "superblock": false, 00:08:23.594 "num_base_bdevs": 3, 00:08:23.594 "num_base_bdevs_discovered": 2, 00:08:23.594 "num_base_bdevs_operational": 3, 00:08:23.594 "base_bdevs_list": [ 00:08:23.594 { 00:08:23.594 "name": "BaseBdev1", 00:08:23.594 "uuid": "4bf8b6e0-624b-434c-a1bf-b08afd490a46", 00:08:23.594 "is_configured": true, 00:08:23.594 "data_offset": 0, 00:08:23.594 "data_size": 65536 00:08:23.594 }, 00:08:23.594 { 00:08:23.594 "name": "BaseBdev2", 00:08:23.594 "uuid": "0e99e2e0-28ca-4064-a592-b4673c31ba4c", 00:08:23.594 "is_configured": true, 00:08:23.594 "data_offset": 0, 00:08:23.594 "data_size": 65536 00:08:23.594 }, 00:08:23.594 { 00:08:23.594 "name": "BaseBdev3", 00:08:23.594 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:23.594 "is_configured": false, 00:08:23.594 "data_offset": 0, 00:08:23.594 "data_size": 0 00:08:23.594 } 00:08:23.594 ] 00:08:23.594 }' 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.594 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.853 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:23.853 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.853 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.853 [2024-09-29 16:46:45.455418] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:23.853 [2024-09-29 16:46:45.455458] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:23.853 [2024-09-29 16:46:45.455471] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:23.854 [2024-09-29 16:46:45.455757] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:23.854 [2024-09-29 16:46:45.455902] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:23.854 [2024-09-29 16:46:45.455912] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:23.854 [2024-09-29 16:46:45.456149] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:23.854 BaseBdev3 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.854 [ 00:08:23.854 { 00:08:23.854 "name": "BaseBdev3", 00:08:23.854 "aliases": [ 00:08:23.854 "6e68d04b-be52-4f41-91d9-75694764dfcf" 00:08:23.854 ], 00:08:23.854 "product_name": "Malloc disk", 00:08:23.854 "block_size": 512, 00:08:23.854 "num_blocks": 65536, 00:08:23.854 "uuid": "6e68d04b-be52-4f41-91d9-75694764dfcf", 00:08:23.854 "assigned_rate_limits": { 00:08:23.854 "rw_ios_per_sec": 0, 00:08:23.854 "rw_mbytes_per_sec": 0, 00:08:23.854 "r_mbytes_per_sec": 0, 00:08:23.854 "w_mbytes_per_sec": 0 00:08:23.854 }, 00:08:23.854 "claimed": true, 00:08:23.854 "claim_type": "exclusive_write", 00:08:23.854 "zoned": false, 00:08:23.854 "supported_io_types": { 00:08:23.854 "read": true, 00:08:23.854 "write": true, 00:08:23.854 "unmap": true, 00:08:23.854 "flush": true, 00:08:23.854 "reset": true, 00:08:23.854 "nvme_admin": false, 00:08:23.854 "nvme_io": false, 00:08:23.854 "nvme_io_md": false, 00:08:23.854 "write_zeroes": true, 00:08:23.854 "zcopy": true, 00:08:23.854 "get_zone_info": false, 00:08:23.854 "zone_management": false, 00:08:23.854 "zone_append": false, 00:08:23.854 "compare": false, 00:08:23.854 "compare_and_write": false, 00:08:23.854 "abort": true, 00:08:23.854 "seek_hole": false, 00:08:23.854 "seek_data": false, 00:08:23.854 "copy": true, 00:08:23.854 "nvme_iov_md": false 00:08:23.854 }, 00:08:23.854 "memory_domains": [ 00:08:23.854 { 00:08:23.854 "dma_device_id": "system", 00:08:23.854 "dma_device_type": 1 00:08:23.854 }, 00:08:23.854 { 00:08:23.854 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:23.854 "dma_device_type": 2 00:08:23.854 } 00:08:23.854 ], 00:08:23.854 "driver_specific": {} 00:08:23.854 } 00:08:23.854 ] 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.854 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.113 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.113 "name": "Existed_Raid", 00:08:24.113 "uuid": "6fdb28a8-ea19-4652-aa7e-6c50ecd88ab2", 00:08:24.113 "strip_size_kb": 64, 00:08:24.113 "state": "online", 00:08:24.113 "raid_level": "concat", 00:08:24.113 "superblock": false, 00:08:24.113 "num_base_bdevs": 3, 00:08:24.113 "num_base_bdevs_discovered": 3, 00:08:24.113 "num_base_bdevs_operational": 3, 00:08:24.113 "base_bdevs_list": [ 00:08:24.113 { 00:08:24.113 "name": "BaseBdev1", 00:08:24.113 "uuid": "4bf8b6e0-624b-434c-a1bf-b08afd490a46", 00:08:24.113 "is_configured": true, 00:08:24.113 "data_offset": 0, 00:08:24.113 "data_size": 65536 00:08:24.113 }, 00:08:24.113 { 00:08:24.113 "name": "BaseBdev2", 00:08:24.113 "uuid": "0e99e2e0-28ca-4064-a592-b4673c31ba4c", 00:08:24.113 "is_configured": true, 00:08:24.113 "data_offset": 0, 00:08:24.113 "data_size": 65536 00:08:24.113 }, 00:08:24.113 { 00:08:24.113 "name": "BaseBdev3", 00:08:24.113 "uuid": "6e68d04b-be52-4f41-91d9-75694764dfcf", 00:08:24.113 "is_configured": true, 00:08:24.113 "data_offset": 0, 00:08:24.113 "data_size": 65536 00:08:24.113 } 00:08:24.113 ] 00:08:24.113 }' 00:08:24.113 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.113 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.372 [2024-09-29 16:46:45.930978] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.372 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:24.372 "name": "Existed_Raid", 00:08:24.372 "aliases": [ 00:08:24.372 "6fdb28a8-ea19-4652-aa7e-6c50ecd88ab2" 00:08:24.372 ], 00:08:24.372 "product_name": "Raid Volume", 00:08:24.372 "block_size": 512, 00:08:24.372 "num_blocks": 196608, 00:08:24.372 "uuid": "6fdb28a8-ea19-4652-aa7e-6c50ecd88ab2", 00:08:24.372 "assigned_rate_limits": { 00:08:24.372 "rw_ios_per_sec": 0, 00:08:24.372 "rw_mbytes_per_sec": 0, 00:08:24.372 "r_mbytes_per_sec": 0, 00:08:24.372 "w_mbytes_per_sec": 0 00:08:24.372 }, 00:08:24.372 "claimed": false, 00:08:24.372 "zoned": false, 00:08:24.372 "supported_io_types": { 00:08:24.372 "read": true, 00:08:24.372 "write": true, 00:08:24.372 "unmap": true, 00:08:24.372 "flush": true, 00:08:24.372 "reset": true, 00:08:24.372 "nvme_admin": false, 00:08:24.372 "nvme_io": false, 00:08:24.372 "nvme_io_md": false, 00:08:24.372 "write_zeroes": true, 00:08:24.372 "zcopy": false, 00:08:24.372 "get_zone_info": false, 00:08:24.372 "zone_management": false, 00:08:24.372 "zone_append": false, 00:08:24.372 "compare": false, 00:08:24.372 "compare_and_write": false, 00:08:24.372 "abort": false, 00:08:24.372 "seek_hole": false, 00:08:24.372 "seek_data": false, 00:08:24.372 "copy": false, 00:08:24.372 "nvme_iov_md": false 00:08:24.372 }, 00:08:24.372 "memory_domains": [ 00:08:24.372 { 00:08:24.372 "dma_device_id": "system", 00:08:24.372 "dma_device_type": 1 00:08:24.372 }, 00:08:24.372 { 00:08:24.372 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.372 "dma_device_type": 2 00:08:24.372 }, 00:08:24.372 { 00:08:24.372 "dma_device_id": "system", 00:08:24.372 "dma_device_type": 1 00:08:24.372 }, 00:08:24.372 { 00:08:24.372 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.372 "dma_device_type": 2 00:08:24.372 }, 00:08:24.372 { 00:08:24.372 "dma_device_id": "system", 00:08:24.373 "dma_device_type": 1 00:08:24.373 }, 00:08:24.373 { 00:08:24.373 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:24.373 "dma_device_type": 2 00:08:24.373 } 00:08:24.373 ], 00:08:24.373 "driver_specific": { 00:08:24.373 "raid": { 00:08:24.373 "uuid": "6fdb28a8-ea19-4652-aa7e-6c50ecd88ab2", 00:08:24.373 "strip_size_kb": 64, 00:08:24.373 "state": "online", 00:08:24.373 "raid_level": "concat", 00:08:24.373 "superblock": false, 00:08:24.373 "num_base_bdevs": 3, 00:08:24.373 "num_base_bdevs_discovered": 3, 00:08:24.373 "num_base_bdevs_operational": 3, 00:08:24.373 "base_bdevs_list": [ 00:08:24.373 { 00:08:24.373 "name": "BaseBdev1", 00:08:24.373 "uuid": "4bf8b6e0-624b-434c-a1bf-b08afd490a46", 00:08:24.373 "is_configured": true, 00:08:24.373 "data_offset": 0, 00:08:24.373 "data_size": 65536 00:08:24.373 }, 00:08:24.373 { 00:08:24.373 "name": "BaseBdev2", 00:08:24.373 "uuid": "0e99e2e0-28ca-4064-a592-b4673c31ba4c", 00:08:24.373 "is_configured": true, 00:08:24.373 "data_offset": 0, 00:08:24.373 "data_size": 65536 00:08:24.373 }, 00:08:24.373 { 00:08:24.373 "name": "BaseBdev3", 00:08:24.373 "uuid": "6e68d04b-be52-4f41-91d9-75694764dfcf", 00:08:24.373 "is_configured": true, 00:08:24.373 "data_offset": 0, 00:08:24.373 "data_size": 65536 00:08:24.373 } 00:08:24.373 ] 00:08:24.373 } 00:08:24.373 } 00:08:24.373 }' 00:08:24.373 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:24.373 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:24.373 BaseBdev2 00:08:24.373 BaseBdev3' 00:08:24.373 16:46:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:24.632 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.633 [2024-09-29 16:46:46.182286] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:24.633 [2024-09-29 16:46:46.182311] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:24.633 [2024-09-29 16:46:46.182360] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.633 "name": "Existed_Raid", 00:08:24.633 "uuid": "6fdb28a8-ea19-4652-aa7e-6c50ecd88ab2", 00:08:24.633 "strip_size_kb": 64, 00:08:24.633 "state": "offline", 00:08:24.633 "raid_level": "concat", 00:08:24.633 "superblock": false, 00:08:24.633 "num_base_bdevs": 3, 00:08:24.633 "num_base_bdevs_discovered": 2, 00:08:24.633 "num_base_bdevs_operational": 2, 00:08:24.633 "base_bdevs_list": [ 00:08:24.633 { 00:08:24.633 "name": null, 00:08:24.633 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.633 "is_configured": false, 00:08:24.633 "data_offset": 0, 00:08:24.633 "data_size": 65536 00:08:24.633 }, 00:08:24.633 { 00:08:24.633 "name": "BaseBdev2", 00:08:24.633 "uuid": "0e99e2e0-28ca-4064-a592-b4673c31ba4c", 00:08:24.633 "is_configured": true, 00:08:24.633 "data_offset": 0, 00:08:24.633 "data_size": 65536 00:08:24.633 }, 00:08:24.633 { 00:08:24.633 "name": "BaseBdev3", 00:08:24.633 "uuid": "6e68d04b-be52-4f41-91d9-75694764dfcf", 00:08:24.633 "is_configured": true, 00:08:24.633 "data_offset": 0, 00:08:24.633 "data_size": 65536 00:08:24.633 } 00:08:24.633 ] 00:08:24.633 }' 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.633 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.201 [2024-09-29 16:46:46.704494] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.201 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.202 [2024-09-29 16:46:46.767648] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:25.202 [2024-09-29 16:46:46.767692] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.202 BaseBdev2 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.202 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.462 [ 00:08:25.462 { 00:08:25.462 "name": "BaseBdev2", 00:08:25.462 "aliases": [ 00:08:25.462 "000a3a38-10cd-4da9-bebe-938b81681563" 00:08:25.462 ], 00:08:25.462 "product_name": "Malloc disk", 00:08:25.462 "block_size": 512, 00:08:25.462 "num_blocks": 65536, 00:08:25.462 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:25.462 "assigned_rate_limits": { 00:08:25.462 "rw_ios_per_sec": 0, 00:08:25.462 "rw_mbytes_per_sec": 0, 00:08:25.462 "r_mbytes_per_sec": 0, 00:08:25.462 "w_mbytes_per_sec": 0 00:08:25.462 }, 00:08:25.462 "claimed": false, 00:08:25.462 "zoned": false, 00:08:25.462 "supported_io_types": { 00:08:25.462 "read": true, 00:08:25.462 "write": true, 00:08:25.462 "unmap": true, 00:08:25.462 "flush": true, 00:08:25.462 "reset": true, 00:08:25.462 "nvme_admin": false, 00:08:25.462 "nvme_io": false, 00:08:25.462 "nvme_io_md": false, 00:08:25.462 "write_zeroes": true, 00:08:25.462 "zcopy": true, 00:08:25.462 "get_zone_info": false, 00:08:25.462 "zone_management": false, 00:08:25.462 "zone_append": false, 00:08:25.462 "compare": false, 00:08:25.462 "compare_and_write": false, 00:08:25.462 "abort": true, 00:08:25.462 "seek_hole": false, 00:08:25.462 "seek_data": false, 00:08:25.462 "copy": true, 00:08:25.462 "nvme_iov_md": false 00:08:25.462 }, 00:08:25.462 "memory_domains": [ 00:08:25.462 { 00:08:25.462 "dma_device_id": "system", 00:08:25.462 "dma_device_type": 1 00:08:25.462 }, 00:08:25.462 { 00:08:25.462 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.462 "dma_device_type": 2 00:08:25.462 } 00:08:25.462 ], 00:08:25.462 "driver_specific": {} 00:08:25.462 } 00:08:25.462 ] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.462 BaseBdev3 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.462 [ 00:08:25.462 { 00:08:25.462 "name": "BaseBdev3", 00:08:25.462 "aliases": [ 00:08:25.462 "777656f5-55b4-4c38-ab44-4e09b1b6c0c9" 00:08:25.462 ], 00:08:25.462 "product_name": "Malloc disk", 00:08:25.462 "block_size": 512, 00:08:25.462 "num_blocks": 65536, 00:08:25.462 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:25.462 "assigned_rate_limits": { 00:08:25.462 "rw_ios_per_sec": 0, 00:08:25.462 "rw_mbytes_per_sec": 0, 00:08:25.462 "r_mbytes_per_sec": 0, 00:08:25.462 "w_mbytes_per_sec": 0 00:08:25.462 }, 00:08:25.462 "claimed": false, 00:08:25.462 "zoned": false, 00:08:25.462 "supported_io_types": { 00:08:25.462 "read": true, 00:08:25.462 "write": true, 00:08:25.462 "unmap": true, 00:08:25.462 "flush": true, 00:08:25.462 "reset": true, 00:08:25.462 "nvme_admin": false, 00:08:25.462 "nvme_io": false, 00:08:25.462 "nvme_io_md": false, 00:08:25.462 "write_zeroes": true, 00:08:25.462 "zcopy": true, 00:08:25.462 "get_zone_info": false, 00:08:25.462 "zone_management": false, 00:08:25.462 "zone_append": false, 00:08:25.462 "compare": false, 00:08:25.462 "compare_and_write": false, 00:08:25.462 "abort": true, 00:08:25.462 "seek_hole": false, 00:08:25.462 "seek_data": false, 00:08:25.462 "copy": true, 00:08:25.462 "nvme_iov_md": false 00:08:25.462 }, 00:08:25.462 "memory_domains": [ 00:08:25.462 { 00:08:25.462 "dma_device_id": "system", 00:08:25.462 "dma_device_type": 1 00:08:25.462 }, 00:08:25.462 { 00:08:25.462 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:25.462 "dma_device_type": 2 00:08:25.462 } 00:08:25.462 ], 00:08:25.462 "driver_specific": {} 00:08:25.462 } 00:08:25.462 ] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.462 [2024-09-29 16:46:46.942099] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:25.462 [2024-09-29 16:46:46.942184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:25.462 [2024-09-29 16:46:46.942224] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:25.462 [2024-09-29 16:46:46.944056] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.462 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.462 "name": "Existed_Raid", 00:08:25.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.462 "strip_size_kb": 64, 00:08:25.462 "state": "configuring", 00:08:25.462 "raid_level": "concat", 00:08:25.462 "superblock": false, 00:08:25.462 "num_base_bdevs": 3, 00:08:25.462 "num_base_bdevs_discovered": 2, 00:08:25.462 "num_base_bdevs_operational": 3, 00:08:25.462 "base_bdevs_list": [ 00:08:25.462 { 00:08:25.462 "name": "BaseBdev1", 00:08:25.462 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.462 "is_configured": false, 00:08:25.462 "data_offset": 0, 00:08:25.462 "data_size": 0 00:08:25.462 }, 00:08:25.462 { 00:08:25.462 "name": "BaseBdev2", 00:08:25.462 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:25.462 "is_configured": true, 00:08:25.462 "data_offset": 0, 00:08:25.462 "data_size": 65536 00:08:25.463 }, 00:08:25.463 { 00:08:25.463 "name": "BaseBdev3", 00:08:25.463 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:25.463 "is_configured": true, 00:08:25.463 "data_offset": 0, 00:08:25.463 "data_size": 65536 00:08:25.463 } 00:08:25.463 ] 00:08:25.463 }' 00:08:25.463 16:46:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.463 16:46:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.722 [2024-09-29 16:46:47.385327] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:25.722 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:25.981 "name": "Existed_Raid", 00:08:25.981 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.981 "strip_size_kb": 64, 00:08:25.981 "state": "configuring", 00:08:25.981 "raid_level": "concat", 00:08:25.981 "superblock": false, 00:08:25.981 "num_base_bdevs": 3, 00:08:25.981 "num_base_bdevs_discovered": 1, 00:08:25.981 "num_base_bdevs_operational": 3, 00:08:25.981 "base_bdevs_list": [ 00:08:25.981 { 00:08:25.981 "name": "BaseBdev1", 00:08:25.981 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:25.981 "is_configured": false, 00:08:25.981 "data_offset": 0, 00:08:25.981 "data_size": 0 00:08:25.981 }, 00:08:25.981 { 00:08:25.981 "name": null, 00:08:25.981 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:25.981 "is_configured": false, 00:08:25.981 "data_offset": 0, 00:08:25.981 "data_size": 65536 00:08:25.981 }, 00:08:25.981 { 00:08:25.981 "name": "BaseBdev3", 00:08:25.981 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:25.981 "is_configured": true, 00:08:25.981 "data_offset": 0, 00:08:25.981 "data_size": 65536 00:08:25.981 } 00:08:25.981 ] 00:08:25.981 }' 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:25.981 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.240 [2024-09-29 16:46:47.843561] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:26.240 BaseBdev1 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.240 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.241 [ 00:08:26.241 { 00:08:26.241 "name": "BaseBdev1", 00:08:26.241 "aliases": [ 00:08:26.241 "e7c54693-7bf6-4749-9915-db64d88b212b" 00:08:26.241 ], 00:08:26.241 "product_name": "Malloc disk", 00:08:26.241 "block_size": 512, 00:08:26.241 "num_blocks": 65536, 00:08:26.241 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:26.241 "assigned_rate_limits": { 00:08:26.241 "rw_ios_per_sec": 0, 00:08:26.241 "rw_mbytes_per_sec": 0, 00:08:26.241 "r_mbytes_per_sec": 0, 00:08:26.241 "w_mbytes_per_sec": 0 00:08:26.241 }, 00:08:26.241 "claimed": true, 00:08:26.241 "claim_type": "exclusive_write", 00:08:26.241 "zoned": false, 00:08:26.241 "supported_io_types": { 00:08:26.241 "read": true, 00:08:26.241 "write": true, 00:08:26.241 "unmap": true, 00:08:26.241 "flush": true, 00:08:26.241 "reset": true, 00:08:26.241 "nvme_admin": false, 00:08:26.241 "nvme_io": false, 00:08:26.241 "nvme_io_md": false, 00:08:26.241 "write_zeroes": true, 00:08:26.241 "zcopy": true, 00:08:26.241 "get_zone_info": false, 00:08:26.241 "zone_management": false, 00:08:26.241 "zone_append": false, 00:08:26.241 "compare": false, 00:08:26.241 "compare_and_write": false, 00:08:26.241 "abort": true, 00:08:26.241 "seek_hole": false, 00:08:26.241 "seek_data": false, 00:08:26.241 "copy": true, 00:08:26.241 "nvme_iov_md": false 00:08:26.241 }, 00:08:26.241 "memory_domains": [ 00:08:26.241 { 00:08:26.241 "dma_device_id": "system", 00:08:26.241 "dma_device_type": 1 00:08:26.241 }, 00:08:26.241 { 00:08:26.241 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:26.241 "dma_device_type": 2 00:08:26.241 } 00:08:26.241 ], 00:08:26.241 "driver_specific": {} 00:08:26.241 } 00:08:26.241 ] 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.241 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.500 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.500 "name": "Existed_Raid", 00:08:26.500 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.500 "strip_size_kb": 64, 00:08:26.500 "state": "configuring", 00:08:26.500 "raid_level": "concat", 00:08:26.500 "superblock": false, 00:08:26.500 "num_base_bdevs": 3, 00:08:26.500 "num_base_bdevs_discovered": 2, 00:08:26.500 "num_base_bdevs_operational": 3, 00:08:26.500 "base_bdevs_list": [ 00:08:26.500 { 00:08:26.500 "name": "BaseBdev1", 00:08:26.500 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:26.500 "is_configured": true, 00:08:26.500 "data_offset": 0, 00:08:26.500 "data_size": 65536 00:08:26.500 }, 00:08:26.500 { 00:08:26.500 "name": null, 00:08:26.500 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:26.500 "is_configured": false, 00:08:26.500 "data_offset": 0, 00:08:26.500 "data_size": 65536 00:08:26.500 }, 00:08:26.500 { 00:08:26.500 "name": "BaseBdev3", 00:08:26.500 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:26.500 "is_configured": true, 00:08:26.500 "data_offset": 0, 00:08:26.500 "data_size": 65536 00:08:26.500 } 00:08:26.500 ] 00:08:26.500 }' 00:08:26.500 16:46:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.500 16:46:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.759 [2024-09-29 16:46:48.394674] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.759 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.760 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.760 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.760 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.760 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.760 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.018 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.018 "name": "Existed_Raid", 00:08:27.018 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.018 "strip_size_kb": 64, 00:08:27.018 "state": "configuring", 00:08:27.018 "raid_level": "concat", 00:08:27.018 "superblock": false, 00:08:27.018 "num_base_bdevs": 3, 00:08:27.018 "num_base_bdevs_discovered": 1, 00:08:27.018 "num_base_bdevs_operational": 3, 00:08:27.018 "base_bdevs_list": [ 00:08:27.018 { 00:08:27.018 "name": "BaseBdev1", 00:08:27.018 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:27.018 "is_configured": true, 00:08:27.018 "data_offset": 0, 00:08:27.018 "data_size": 65536 00:08:27.018 }, 00:08:27.018 { 00:08:27.018 "name": null, 00:08:27.018 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:27.018 "is_configured": false, 00:08:27.018 "data_offset": 0, 00:08:27.019 "data_size": 65536 00:08:27.019 }, 00:08:27.019 { 00:08:27.019 "name": null, 00:08:27.019 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:27.019 "is_configured": false, 00:08:27.019 "data_offset": 0, 00:08:27.019 "data_size": 65536 00:08:27.019 } 00:08:27.019 ] 00:08:27.019 }' 00:08:27.019 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.019 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.277 [2024-09-29 16:46:48.881870] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.277 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.277 "name": "Existed_Raid", 00:08:27.277 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.277 "strip_size_kb": 64, 00:08:27.277 "state": "configuring", 00:08:27.277 "raid_level": "concat", 00:08:27.277 "superblock": false, 00:08:27.277 "num_base_bdevs": 3, 00:08:27.277 "num_base_bdevs_discovered": 2, 00:08:27.277 "num_base_bdevs_operational": 3, 00:08:27.277 "base_bdevs_list": [ 00:08:27.277 { 00:08:27.277 "name": "BaseBdev1", 00:08:27.278 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:27.278 "is_configured": true, 00:08:27.278 "data_offset": 0, 00:08:27.278 "data_size": 65536 00:08:27.278 }, 00:08:27.278 { 00:08:27.278 "name": null, 00:08:27.278 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:27.278 "is_configured": false, 00:08:27.278 "data_offset": 0, 00:08:27.278 "data_size": 65536 00:08:27.278 }, 00:08:27.278 { 00:08:27.278 "name": "BaseBdev3", 00:08:27.278 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:27.278 "is_configured": true, 00:08:27.278 "data_offset": 0, 00:08:27.278 "data_size": 65536 00:08:27.278 } 00:08:27.278 ] 00:08:27.278 }' 00:08:27.278 16:46:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.278 16:46:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.845 [2024-09-29 16:46:49.393043] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.845 "name": "Existed_Raid", 00:08:27.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.845 "strip_size_kb": 64, 00:08:27.845 "state": "configuring", 00:08:27.845 "raid_level": "concat", 00:08:27.845 "superblock": false, 00:08:27.845 "num_base_bdevs": 3, 00:08:27.845 "num_base_bdevs_discovered": 1, 00:08:27.845 "num_base_bdevs_operational": 3, 00:08:27.845 "base_bdevs_list": [ 00:08:27.845 { 00:08:27.845 "name": null, 00:08:27.845 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:27.845 "is_configured": false, 00:08:27.845 "data_offset": 0, 00:08:27.845 "data_size": 65536 00:08:27.845 }, 00:08:27.845 { 00:08:27.845 "name": null, 00:08:27.845 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:27.845 "is_configured": false, 00:08:27.845 "data_offset": 0, 00:08:27.845 "data_size": 65536 00:08:27.845 }, 00:08:27.845 { 00:08:27.845 "name": "BaseBdev3", 00:08:27.845 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:27.845 "is_configured": true, 00:08:27.845 "data_offset": 0, 00:08:27.845 "data_size": 65536 00:08:27.845 } 00:08:27.845 ] 00:08:27.845 }' 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.845 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.413 [2024-09-29 16:46:49.878574] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.413 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.414 "name": "Existed_Raid", 00:08:28.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.414 "strip_size_kb": 64, 00:08:28.414 "state": "configuring", 00:08:28.414 "raid_level": "concat", 00:08:28.414 "superblock": false, 00:08:28.414 "num_base_bdevs": 3, 00:08:28.414 "num_base_bdevs_discovered": 2, 00:08:28.414 "num_base_bdevs_operational": 3, 00:08:28.414 "base_bdevs_list": [ 00:08:28.414 { 00:08:28.414 "name": null, 00:08:28.414 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:28.414 "is_configured": false, 00:08:28.414 "data_offset": 0, 00:08:28.414 "data_size": 65536 00:08:28.414 }, 00:08:28.414 { 00:08:28.414 "name": "BaseBdev2", 00:08:28.414 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:28.414 "is_configured": true, 00:08:28.414 "data_offset": 0, 00:08:28.414 "data_size": 65536 00:08:28.414 }, 00:08:28.414 { 00:08:28.414 "name": "BaseBdev3", 00:08:28.414 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:28.414 "is_configured": true, 00:08:28.414 "data_offset": 0, 00:08:28.414 "data_size": 65536 00:08:28.414 } 00:08:28.414 ] 00:08:28.414 }' 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.414 16:46:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.673 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e7c54693-7bf6-4749-9915-db64d88b212b 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.933 [2024-09-29 16:46:50.380870] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:28.933 [2024-09-29 16:46:50.380908] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:28.933 [2024-09-29 16:46:50.380917] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:28.933 [2024-09-29 16:46:50.381156] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:28.933 [2024-09-29 16:46:50.381272] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:28.933 [2024-09-29 16:46:50.381281] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:28.933 [2024-09-29 16:46:50.381460] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:28.933 NewBaseBdev 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:28.933 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.934 [ 00:08:28.934 { 00:08:28.934 "name": "NewBaseBdev", 00:08:28.934 "aliases": [ 00:08:28.934 "e7c54693-7bf6-4749-9915-db64d88b212b" 00:08:28.934 ], 00:08:28.934 "product_name": "Malloc disk", 00:08:28.934 "block_size": 512, 00:08:28.934 "num_blocks": 65536, 00:08:28.934 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:28.934 "assigned_rate_limits": { 00:08:28.934 "rw_ios_per_sec": 0, 00:08:28.934 "rw_mbytes_per_sec": 0, 00:08:28.934 "r_mbytes_per_sec": 0, 00:08:28.934 "w_mbytes_per_sec": 0 00:08:28.934 }, 00:08:28.934 "claimed": true, 00:08:28.934 "claim_type": "exclusive_write", 00:08:28.934 "zoned": false, 00:08:28.934 "supported_io_types": { 00:08:28.934 "read": true, 00:08:28.934 "write": true, 00:08:28.934 "unmap": true, 00:08:28.934 "flush": true, 00:08:28.934 "reset": true, 00:08:28.934 "nvme_admin": false, 00:08:28.934 "nvme_io": false, 00:08:28.934 "nvme_io_md": false, 00:08:28.934 "write_zeroes": true, 00:08:28.934 "zcopy": true, 00:08:28.934 "get_zone_info": false, 00:08:28.934 "zone_management": false, 00:08:28.934 "zone_append": false, 00:08:28.934 "compare": false, 00:08:28.934 "compare_and_write": false, 00:08:28.934 "abort": true, 00:08:28.934 "seek_hole": false, 00:08:28.934 "seek_data": false, 00:08:28.934 "copy": true, 00:08:28.934 "nvme_iov_md": false 00:08:28.934 }, 00:08:28.934 "memory_domains": [ 00:08:28.934 { 00:08:28.934 "dma_device_id": "system", 00:08:28.934 "dma_device_type": 1 00:08:28.934 }, 00:08:28.934 { 00:08:28.934 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.934 "dma_device_type": 2 00:08:28.934 } 00:08:28.934 ], 00:08:28.934 "driver_specific": {} 00:08:28.934 } 00:08:28.934 ] 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.934 "name": "Existed_Raid", 00:08:28.934 "uuid": "de7f7e5c-9637-4028-af70-13335b98f2e2", 00:08:28.934 "strip_size_kb": 64, 00:08:28.934 "state": "online", 00:08:28.934 "raid_level": "concat", 00:08:28.934 "superblock": false, 00:08:28.934 "num_base_bdevs": 3, 00:08:28.934 "num_base_bdevs_discovered": 3, 00:08:28.934 "num_base_bdevs_operational": 3, 00:08:28.934 "base_bdevs_list": [ 00:08:28.934 { 00:08:28.934 "name": "NewBaseBdev", 00:08:28.934 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:28.934 "is_configured": true, 00:08:28.934 "data_offset": 0, 00:08:28.934 "data_size": 65536 00:08:28.934 }, 00:08:28.934 { 00:08:28.934 "name": "BaseBdev2", 00:08:28.934 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:28.934 "is_configured": true, 00:08:28.934 "data_offset": 0, 00:08:28.934 "data_size": 65536 00:08:28.934 }, 00:08:28.934 { 00:08:28.934 "name": "BaseBdev3", 00:08:28.934 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:28.934 "is_configured": true, 00:08:28.934 "data_offset": 0, 00:08:28.934 "data_size": 65536 00:08:28.934 } 00:08:28.934 ] 00:08:28.934 }' 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.934 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.194 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:29.194 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:29.194 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:29.194 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:29.194 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:29.194 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:29.453 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:29.453 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.453 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.453 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:29.453 [2024-09-29 16:46:50.872375] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:29.453 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.453 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:29.453 "name": "Existed_Raid", 00:08:29.454 "aliases": [ 00:08:29.454 "de7f7e5c-9637-4028-af70-13335b98f2e2" 00:08:29.454 ], 00:08:29.454 "product_name": "Raid Volume", 00:08:29.454 "block_size": 512, 00:08:29.454 "num_blocks": 196608, 00:08:29.454 "uuid": "de7f7e5c-9637-4028-af70-13335b98f2e2", 00:08:29.454 "assigned_rate_limits": { 00:08:29.454 "rw_ios_per_sec": 0, 00:08:29.454 "rw_mbytes_per_sec": 0, 00:08:29.454 "r_mbytes_per_sec": 0, 00:08:29.454 "w_mbytes_per_sec": 0 00:08:29.454 }, 00:08:29.454 "claimed": false, 00:08:29.454 "zoned": false, 00:08:29.454 "supported_io_types": { 00:08:29.454 "read": true, 00:08:29.454 "write": true, 00:08:29.454 "unmap": true, 00:08:29.454 "flush": true, 00:08:29.454 "reset": true, 00:08:29.454 "nvme_admin": false, 00:08:29.454 "nvme_io": false, 00:08:29.454 "nvme_io_md": false, 00:08:29.454 "write_zeroes": true, 00:08:29.454 "zcopy": false, 00:08:29.454 "get_zone_info": false, 00:08:29.454 "zone_management": false, 00:08:29.454 "zone_append": false, 00:08:29.454 "compare": false, 00:08:29.454 "compare_and_write": false, 00:08:29.454 "abort": false, 00:08:29.454 "seek_hole": false, 00:08:29.454 "seek_data": false, 00:08:29.454 "copy": false, 00:08:29.454 "nvme_iov_md": false 00:08:29.454 }, 00:08:29.454 "memory_domains": [ 00:08:29.454 { 00:08:29.454 "dma_device_id": "system", 00:08:29.454 "dma_device_type": 1 00:08:29.454 }, 00:08:29.454 { 00:08:29.454 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.454 "dma_device_type": 2 00:08:29.454 }, 00:08:29.454 { 00:08:29.454 "dma_device_id": "system", 00:08:29.454 "dma_device_type": 1 00:08:29.454 }, 00:08:29.454 { 00:08:29.454 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.454 "dma_device_type": 2 00:08:29.454 }, 00:08:29.454 { 00:08:29.454 "dma_device_id": "system", 00:08:29.454 "dma_device_type": 1 00:08:29.454 }, 00:08:29.454 { 00:08:29.454 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:29.454 "dma_device_type": 2 00:08:29.454 } 00:08:29.454 ], 00:08:29.454 "driver_specific": { 00:08:29.454 "raid": { 00:08:29.454 "uuid": "de7f7e5c-9637-4028-af70-13335b98f2e2", 00:08:29.454 "strip_size_kb": 64, 00:08:29.454 "state": "online", 00:08:29.454 "raid_level": "concat", 00:08:29.454 "superblock": false, 00:08:29.454 "num_base_bdevs": 3, 00:08:29.454 "num_base_bdevs_discovered": 3, 00:08:29.454 "num_base_bdevs_operational": 3, 00:08:29.454 "base_bdevs_list": [ 00:08:29.454 { 00:08:29.454 "name": "NewBaseBdev", 00:08:29.454 "uuid": "e7c54693-7bf6-4749-9915-db64d88b212b", 00:08:29.454 "is_configured": true, 00:08:29.454 "data_offset": 0, 00:08:29.454 "data_size": 65536 00:08:29.454 }, 00:08:29.454 { 00:08:29.454 "name": "BaseBdev2", 00:08:29.454 "uuid": "000a3a38-10cd-4da9-bebe-938b81681563", 00:08:29.454 "is_configured": true, 00:08:29.454 "data_offset": 0, 00:08:29.454 "data_size": 65536 00:08:29.454 }, 00:08:29.454 { 00:08:29.454 "name": "BaseBdev3", 00:08:29.454 "uuid": "777656f5-55b4-4c38-ab44-4e09b1b6c0c9", 00:08:29.454 "is_configured": true, 00:08:29.454 "data_offset": 0, 00:08:29.454 "data_size": 65536 00:08:29.454 } 00:08:29.454 ] 00:08:29.454 } 00:08:29.454 } 00:08:29.454 }' 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:29.454 BaseBdev2 00:08:29.454 BaseBdev3' 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.454 16:46:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.454 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.715 [2024-09-29 16:46:51.143580] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:29.715 [2024-09-29 16:46:51.143607] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:29.715 [2024-09-29 16:46:51.143671] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:29.715 [2024-09-29 16:46:51.143772] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:29.715 [2024-09-29 16:46:51.143788] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76446 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 76446 ']' 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 76446 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76446 00:08:29.715 killing process with pid 76446 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76446' 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 76446 00:08:29.715 [2024-09-29 16:46:51.193673] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:29.715 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 76446 00:08:29.715 [2024-09-29 16:46:51.224054] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:29.975 ************************************ 00:08:29.975 END TEST raid_state_function_test 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:29.975 00:08:29.975 real 0m8.853s 00:08:29.975 user 0m15.148s 00:08:29.975 sys 0m1.766s 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.975 ************************************ 00:08:29.975 16:46:51 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:08:29.975 16:46:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:29.975 16:46:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:29.975 16:46:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:29.975 ************************************ 00:08:29.975 START TEST raid_state_function_test_sb 00:08:29.975 ************************************ 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 true 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=77051 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77051' 00:08:29.975 Process raid pid: 77051 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 77051 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 77051 ']' 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:29.975 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:29.975 16:46:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:29.975 [2024-09-29 16:46:51.623326] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:29.975 [2024-09-29 16:46:51.623533] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:30.243 [2024-09-29 16:46:51.749764] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:30.243 [2024-09-29 16:46:51.795695] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:30.243 [2024-09-29 16:46:51.837295] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:30.243 [2024-09-29 16:46:51.837409] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:30.819 [2024-09-29 16:46:52.466385] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:30.819 [2024-09-29 16:46:52.466433] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:30.819 [2024-09-29 16:46:52.466451] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:30.819 [2024-09-29 16:46:52.466462] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:30.819 [2024-09-29 16:46:52.466469] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:30.819 [2024-09-29 16:46:52.466479] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.819 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.079 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.079 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.079 "name": "Existed_Raid", 00:08:31.079 "uuid": "67735f34-d095-48fd-b67c-b41917763b98", 00:08:31.079 "strip_size_kb": 64, 00:08:31.079 "state": "configuring", 00:08:31.079 "raid_level": "concat", 00:08:31.079 "superblock": true, 00:08:31.079 "num_base_bdevs": 3, 00:08:31.079 "num_base_bdevs_discovered": 0, 00:08:31.079 "num_base_bdevs_operational": 3, 00:08:31.079 "base_bdevs_list": [ 00:08:31.079 { 00:08:31.079 "name": "BaseBdev1", 00:08:31.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.079 "is_configured": false, 00:08:31.079 "data_offset": 0, 00:08:31.079 "data_size": 0 00:08:31.079 }, 00:08:31.079 { 00:08:31.079 "name": "BaseBdev2", 00:08:31.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.079 "is_configured": false, 00:08:31.079 "data_offset": 0, 00:08:31.079 "data_size": 0 00:08:31.079 }, 00:08:31.079 { 00:08:31.079 "name": "BaseBdev3", 00:08:31.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.079 "is_configured": false, 00:08:31.079 "data_offset": 0, 00:08:31.079 "data_size": 0 00:08:31.079 } 00:08:31.079 ] 00:08:31.079 }' 00:08:31.079 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.079 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.339 [2024-09-29 16:46:52.953407] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:31.339 [2024-09-29 16:46:52.953455] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.339 [2024-09-29 16:46:52.965404] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:31.339 [2024-09-29 16:46:52.965446] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:31.339 [2024-09-29 16:46:52.965454] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:31.339 [2024-09-29 16:46:52.965463] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:31.339 [2024-09-29 16:46:52.965469] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:31.339 [2024-09-29 16:46:52.965477] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.339 [2024-09-29 16:46:52.986148] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:31.339 BaseBdev1 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.339 16:46:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.599 [ 00:08:31.599 { 00:08:31.599 "name": "BaseBdev1", 00:08:31.599 "aliases": [ 00:08:31.599 "3fec4c9c-7512-42a3-ac25-bd67189b96be" 00:08:31.599 ], 00:08:31.599 "product_name": "Malloc disk", 00:08:31.599 "block_size": 512, 00:08:31.599 "num_blocks": 65536, 00:08:31.599 "uuid": "3fec4c9c-7512-42a3-ac25-bd67189b96be", 00:08:31.599 "assigned_rate_limits": { 00:08:31.599 "rw_ios_per_sec": 0, 00:08:31.599 "rw_mbytes_per_sec": 0, 00:08:31.599 "r_mbytes_per_sec": 0, 00:08:31.599 "w_mbytes_per_sec": 0 00:08:31.599 }, 00:08:31.599 "claimed": true, 00:08:31.599 "claim_type": "exclusive_write", 00:08:31.599 "zoned": false, 00:08:31.599 "supported_io_types": { 00:08:31.600 "read": true, 00:08:31.600 "write": true, 00:08:31.600 "unmap": true, 00:08:31.600 "flush": true, 00:08:31.600 "reset": true, 00:08:31.600 "nvme_admin": false, 00:08:31.600 "nvme_io": false, 00:08:31.600 "nvme_io_md": false, 00:08:31.600 "write_zeroes": true, 00:08:31.600 "zcopy": true, 00:08:31.600 "get_zone_info": false, 00:08:31.600 "zone_management": false, 00:08:31.600 "zone_append": false, 00:08:31.600 "compare": false, 00:08:31.600 "compare_and_write": false, 00:08:31.600 "abort": true, 00:08:31.600 "seek_hole": false, 00:08:31.600 "seek_data": false, 00:08:31.600 "copy": true, 00:08:31.600 "nvme_iov_md": false 00:08:31.600 }, 00:08:31.600 "memory_domains": [ 00:08:31.600 { 00:08:31.600 "dma_device_id": "system", 00:08:31.600 "dma_device_type": 1 00:08:31.600 }, 00:08:31.600 { 00:08:31.600 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:31.600 "dma_device_type": 2 00:08:31.600 } 00:08:31.600 ], 00:08:31.600 "driver_specific": {} 00:08:31.600 } 00:08:31.600 ] 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.600 "name": "Existed_Raid", 00:08:31.600 "uuid": "09100fe8-0269-45fc-aa51-b536b4b8cbf4", 00:08:31.600 "strip_size_kb": 64, 00:08:31.600 "state": "configuring", 00:08:31.600 "raid_level": "concat", 00:08:31.600 "superblock": true, 00:08:31.600 "num_base_bdevs": 3, 00:08:31.600 "num_base_bdevs_discovered": 1, 00:08:31.600 "num_base_bdevs_operational": 3, 00:08:31.600 "base_bdevs_list": [ 00:08:31.600 { 00:08:31.600 "name": "BaseBdev1", 00:08:31.600 "uuid": "3fec4c9c-7512-42a3-ac25-bd67189b96be", 00:08:31.600 "is_configured": true, 00:08:31.600 "data_offset": 2048, 00:08:31.600 "data_size": 63488 00:08:31.600 }, 00:08:31.600 { 00:08:31.600 "name": "BaseBdev2", 00:08:31.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.600 "is_configured": false, 00:08:31.600 "data_offset": 0, 00:08:31.600 "data_size": 0 00:08:31.600 }, 00:08:31.600 { 00:08:31.600 "name": "BaseBdev3", 00:08:31.600 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.600 "is_configured": false, 00:08:31.600 "data_offset": 0, 00:08:31.600 "data_size": 0 00:08:31.600 } 00:08:31.600 ] 00:08:31.600 }' 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.600 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.859 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:31.859 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.859 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.859 [2024-09-29 16:46:53.453372] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:31.859 [2024-09-29 16:46:53.453460] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:31.859 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.859 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:31.859 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.860 [2024-09-29 16:46:53.465423] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:31.860 [2024-09-29 16:46:53.467331] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:31.860 [2024-09-29 16:46:53.467375] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:31.860 [2024-09-29 16:46:53.467385] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:31.860 [2024-09-29 16:46:53.467395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.860 "name": "Existed_Raid", 00:08:31.860 "uuid": "f4e43ad5-8c3b-4691-a6d0-3458da5579cb", 00:08:31.860 "strip_size_kb": 64, 00:08:31.860 "state": "configuring", 00:08:31.860 "raid_level": "concat", 00:08:31.860 "superblock": true, 00:08:31.860 "num_base_bdevs": 3, 00:08:31.860 "num_base_bdevs_discovered": 1, 00:08:31.860 "num_base_bdevs_operational": 3, 00:08:31.860 "base_bdevs_list": [ 00:08:31.860 { 00:08:31.860 "name": "BaseBdev1", 00:08:31.860 "uuid": "3fec4c9c-7512-42a3-ac25-bd67189b96be", 00:08:31.860 "is_configured": true, 00:08:31.860 "data_offset": 2048, 00:08:31.860 "data_size": 63488 00:08:31.860 }, 00:08:31.860 { 00:08:31.860 "name": "BaseBdev2", 00:08:31.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.860 "is_configured": false, 00:08:31.860 "data_offset": 0, 00:08:31.860 "data_size": 0 00:08:31.860 }, 00:08:31.860 { 00:08:31.860 "name": "BaseBdev3", 00:08:31.860 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.860 "is_configured": false, 00:08:31.860 "data_offset": 0, 00:08:31.860 "data_size": 0 00:08:31.860 } 00:08:31.860 ] 00:08:31.860 }' 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.860 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.429 [2024-09-29 16:46:53.891890] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:32.429 BaseBdev2 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.429 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.429 [ 00:08:32.429 { 00:08:32.429 "name": "BaseBdev2", 00:08:32.429 "aliases": [ 00:08:32.429 "0f2b4ad9-50dc-44fc-b2fe-1b3eb7bbeff9" 00:08:32.429 ], 00:08:32.429 "product_name": "Malloc disk", 00:08:32.429 "block_size": 512, 00:08:32.429 "num_blocks": 65536, 00:08:32.429 "uuid": "0f2b4ad9-50dc-44fc-b2fe-1b3eb7bbeff9", 00:08:32.429 "assigned_rate_limits": { 00:08:32.429 "rw_ios_per_sec": 0, 00:08:32.429 "rw_mbytes_per_sec": 0, 00:08:32.429 "r_mbytes_per_sec": 0, 00:08:32.429 "w_mbytes_per_sec": 0 00:08:32.429 }, 00:08:32.429 "claimed": true, 00:08:32.430 "claim_type": "exclusive_write", 00:08:32.430 "zoned": false, 00:08:32.430 "supported_io_types": { 00:08:32.430 "read": true, 00:08:32.430 "write": true, 00:08:32.430 "unmap": true, 00:08:32.430 "flush": true, 00:08:32.430 "reset": true, 00:08:32.430 "nvme_admin": false, 00:08:32.430 "nvme_io": false, 00:08:32.430 "nvme_io_md": false, 00:08:32.430 "write_zeroes": true, 00:08:32.430 "zcopy": true, 00:08:32.430 "get_zone_info": false, 00:08:32.430 "zone_management": false, 00:08:32.430 "zone_append": false, 00:08:32.430 "compare": false, 00:08:32.430 "compare_and_write": false, 00:08:32.430 "abort": true, 00:08:32.430 "seek_hole": false, 00:08:32.430 "seek_data": false, 00:08:32.430 "copy": true, 00:08:32.430 "nvme_iov_md": false 00:08:32.430 }, 00:08:32.430 "memory_domains": [ 00:08:32.430 { 00:08:32.430 "dma_device_id": "system", 00:08:32.430 "dma_device_type": 1 00:08:32.430 }, 00:08:32.430 { 00:08:32.430 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:32.430 "dma_device_type": 2 00:08:32.430 } 00:08:32.430 ], 00:08:32.430 "driver_specific": {} 00:08:32.430 } 00:08:32.430 ] 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:32.430 "name": "Existed_Raid", 00:08:32.430 "uuid": "f4e43ad5-8c3b-4691-a6d0-3458da5579cb", 00:08:32.430 "strip_size_kb": 64, 00:08:32.430 "state": "configuring", 00:08:32.430 "raid_level": "concat", 00:08:32.430 "superblock": true, 00:08:32.430 "num_base_bdevs": 3, 00:08:32.430 "num_base_bdevs_discovered": 2, 00:08:32.430 "num_base_bdevs_operational": 3, 00:08:32.430 "base_bdevs_list": [ 00:08:32.430 { 00:08:32.430 "name": "BaseBdev1", 00:08:32.430 "uuid": "3fec4c9c-7512-42a3-ac25-bd67189b96be", 00:08:32.430 "is_configured": true, 00:08:32.430 "data_offset": 2048, 00:08:32.430 "data_size": 63488 00:08:32.430 }, 00:08:32.430 { 00:08:32.430 "name": "BaseBdev2", 00:08:32.430 "uuid": "0f2b4ad9-50dc-44fc-b2fe-1b3eb7bbeff9", 00:08:32.430 "is_configured": true, 00:08:32.430 "data_offset": 2048, 00:08:32.430 "data_size": 63488 00:08:32.430 }, 00:08:32.430 { 00:08:32.430 "name": "BaseBdev3", 00:08:32.430 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:32.430 "is_configured": false, 00:08:32.430 "data_offset": 0, 00:08:32.430 "data_size": 0 00:08:32.430 } 00:08:32.430 ] 00:08:32.430 }' 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:32.430 16:46:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:32.999 [2024-09-29 16:46:54.382009] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:32.999 BaseBdev3 00:08:32.999 [2024-09-29 16:46:54.382318] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:32.999 [2024-09-29 16:46:54.382349] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:32.999 [2024-09-29 16:46:54.382628] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:32.999 [2024-09-29 16:46:54.382794] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:32.999 [2024-09-29 16:46:54.382810] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:32.999 [2024-09-29 16:46:54.382930] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:32.999 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.000 [ 00:08:33.000 { 00:08:33.000 "name": "BaseBdev3", 00:08:33.000 "aliases": [ 00:08:33.000 "2c8a4419-577a-46c7-95e6-3586e99e7865" 00:08:33.000 ], 00:08:33.000 "product_name": "Malloc disk", 00:08:33.000 "block_size": 512, 00:08:33.000 "num_blocks": 65536, 00:08:33.000 "uuid": "2c8a4419-577a-46c7-95e6-3586e99e7865", 00:08:33.000 "assigned_rate_limits": { 00:08:33.000 "rw_ios_per_sec": 0, 00:08:33.000 "rw_mbytes_per_sec": 0, 00:08:33.000 "r_mbytes_per_sec": 0, 00:08:33.000 "w_mbytes_per_sec": 0 00:08:33.000 }, 00:08:33.000 "claimed": true, 00:08:33.000 "claim_type": "exclusive_write", 00:08:33.000 "zoned": false, 00:08:33.000 "supported_io_types": { 00:08:33.000 "read": true, 00:08:33.000 "write": true, 00:08:33.000 "unmap": true, 00:08:33.000 "flush": true, 00:08:33.000 "reset": true, 00:08:33.000 "nvme_admin": false, 00:08:33.000 "nvme_io": false, 00:08:33.000 "nvme_io_md": false, 00:08:33.000 "write_zeroes": true, 00:08:33.000 "zcopy": true, 00:08:33.000 "get_zone_info": false, 00:08:33.000 "zone_management": false, 00:08:33.000 "zone_append": false, 00:08:33.000 "compare": false, 00:08:33.000 "compare_and_write": false, 00:08:33.000 "abort": true, 00:08:33.000 "seek_hole": false, 00:08:33.000 "seek_data": false, 00:08:33.000 "copy": true, 00:08:33.000 "nvme_iov_md": false 00:08:33.000 }, 00:08:33.000 "memory_domains": [ 00:08:33.000 { 00:08:33.000 "dma_device_id": "system", 00:08:33.000 "dma_device_type": 1 00:08:33.000 }, 00:08:33.000 { 00:08:33.000 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.000 "dma_device_type": 2 00:08:33.000 } 00:08:33.000 ], 00:08:33.000 "driver_specific": {} 00:08:33.000 } 00:08:33.000 ] 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.000 "name": "Existed_Raid", 00:08:33.000 "uuid": "f4e43ad5-8c3b-4691-a6d0-3458da5579cb", 00:08:33.000 "strip_size_kb": 64, 00:08:33.000 "state": "online", 00:08:33.000 "raid_level": "concat", 00:08:33.000 "superblock": true, 00:08:33.000 "num_base_bdevs": 3, 00:08:33.000 "num_base_bdevs_discovered": 3, 00:08:33.000 "num_base_bdevs_operational": 3, 00:08:33.000 "base_bdevs_list": [ 00:08:33.000 { 00:08:33.000 "name": "BaseBdev1", 00:08:33.000 "uuid": "3fec4c9c-7512-42a3-ac25-bd67189b96be", 00:08:33.000 "is_configured": true, 00:08:33.000 "data_offset": 2048, 00:08:33.000 "data_size": 63488 00:08:33.000 }, 00:08:33.000 { 00:08:33.000 "name": "BaseBdev2", 00:08:33.000 "uuid": "0f2b4ad9-50dc-44fc-b2fe-1b3eb7bbeff9", 00:08:33.000 "is_configured": true, 00:08:33.000 "data_offset": 2048, 00:08:33.000 "data_size": 63488 00:08:33.000 }, 00:08:33.000 { 00:08:33.000 "name": "BaseBdev3", 00:08:33.000 "uuid": "2c8a4419-577a-46c7-95e6-3586e99e7865", 00:08:33.000 "is_configured": true, 00:08:33.000 "data_offset": 2048, 00:08:33.000 "data_size": 63488 00:08:33.000 } 00:08:33.000 ] 00:08:33.000 }' 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.000 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:33.260 [2024-09-29 16:46:54.877462] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.260 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:33.260 "name": "Existed_Raid", 00:08:33.260 "aliases": [ 00:08:33.260 "f4e43ad5-8c3b-4691-a6d0-3458da5579cb" 00:08:33.260 ], 00:08:33.260 "product_name": "Raid Volume", 00:08:33.260 "block_size": 512, 00:08:33.260 "num_blocks": 190464, 00:08:33.260 "uuid": "f4e43ad5-8c3b-4691-a6d0-3458da5579cb", 00:08:33.260 "assigned_rate_limits": { 00:08:33.260 "rw_ios_per_sec": 0, 00:08:33.260 "rw_mbytes_per_sec": 0, 00:08:33.260 "r_mbytes_per_sec": 0, 00:08:33.260 "w_mbytes_per_sec": 0 00:08:33.260 }, 00:08:33.260 "claimed": false, 00:08:33.260 "zoned": false, 00:08:33.260 "supported_io_types": { 00:08:33.260 "read": true, 00:08:33.260 "write": true, 00:08:33.260 "unmap": true, 00:08:33.260 "flush": true, 00:08:33.260 "reset": true, 00:08:33.260 "nvme_admin": false, 00:08:33.260 "nvme_io": false, 00:08:33.260 "nvme_io_md": false, 00:08:33.260 "write_zeroes": true, 00:08:33.260 "zcopy": false, 00:08:33.260 "get_zone_info": false, 00:08:33.260 "zone_management": false, 00:08:33.260 "zone_append": false, 00:08:33.260 "compare": false, 00:08:33.260 "compare_and_write": false, 00:08:33.260 "abort": false, 00:08:33.260 "seek_hole": false, 00:08:33.260 "seek_data": false, 00:08:33.260 "copy": false, 00:08:33.260 "nvme_iov_md": false 00:08:33.260 }, 00:08:33.260 "memory_domains": [ 00:08:33.260 { 00:08:33.260 "dma_device_id": "system", 00:08:33.260 "dma_device_type": 1 00:08:33.260 }, 00:08:33.260 { 00:08:33.260 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.260 "dma_device_type": 2 00:08:33.261 }, 00:08:33.261 { 00:08:33.261 "dma_device_id": "system", 00:08:33.261 "dma_device_type": 1 00:08:33.261 }, 00:08:33.261 { 00:08:33.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.261 "dma_device_type": 2 00:08:33.261 }, 00:08:33.261 { 00:08:33.261 "dma_device_id": "system", 00:08:33.261 "dma_device_type": 1 00:08:33.261 }, 00:08:33.261 { 00:08:33.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.261 "dma_device_type": 2 00:08:33.261 } 00:08:33.261 ], 00:08:33.261 "driver_specific": { 00:08:33.261 "raid": { 00:08:33.261 "uuid": "f4e43ad5-8c3b-4691-a6d0-3458da5579cb", 00:08:33.261 "strip_size_kb": 64, 00:08:33.261 "state": "online", 00:08:33.261 "raid_level": "concat", 00:08:33.261 "superblock": true, 00:08:33.261 "num_base_bdevs": 3, 00:08:33.261 "num_base_bdevs_discovered": 3, 00:08:33.261 "num_base_bdevs_operational": 3, 00:08:33.261 "base_bdevs_list": [ 00:08:33.261 { 00:08:33.261 "name": "BaseBdev1", 00:08:33.261 "uuid": "3fec4c9c-7512-42a3-ac25-bd67189b96be", 00:08:33.261 "is_configured": true, 00:08:33.261 "data_offset": 2048, 00:08:33.261 "data_size": 63488 00:08:33.261 }, 00:08:33.261 { 00:08:33.261 "name": "BaseBdev2", 00:08:33.261 "uuid": "0f2b4ad9-50dc-44fc-b2fe-1b3eb7bbeff9", 00:08:33.261 "is_configured": true, 00:08:33.261 "data_offset": 2048, 00:08:33.261 "data_size": 63488 00:08:33.261 }, 00:08:33.261 { 00:08:33.261 "name": "BaseBdev3", 00:08:33.261 "uuid": "2c8a4419-577a-46c7-95e6-3586e99e7865", 00:08:33.261 "is_configured": true, 00:08:33.261 "data_offset": 2048, 00:08:33.261 "data_size": 63488 00:08:33.261 } 00:08:33.261 ] 00:08:33.261 } 00:08:33.261 } 00:08:33.261 }' 00:08:33.261 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:33.521 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:33.521 BaseBdev2 00:08:33.521 BaseBdev3' 00:08:33.521 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:33.521 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:33.522 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:33.522 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:33.522 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.522 16:46:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.522 16:46:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.522 [2024-09-29 16:46:55.144792] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:33.522 [2024-09-29 16:46:55.144856] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:33.522 [2024-09-29 16:46:55.144908] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:33.522 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.782 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.782 "name": "Existed_Raid", 00:08:33.782 "uuid": "f4e43ad5-8c3b-4691-a6d0-3458da5579cb", 00:08:33.782 "strip_size_kb": 64, 00:08:33.782 "state": "offline", 00:08:33.782 "raid_level": "concat", 00:08:33.782 "superblock": true, 00:08:33.782 "num_base_bdevs": 3, 00:08:33.782 "num_base_bdevs_discovered": 2, 00:08:33.782 "num_base_bdevs_operational": 2, 00:08:33.782 "base_bdevs_list": [ 00:08:33.782 { 00:08:33.782 "name": null, 00:08:33.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:33.782 "is_configured": false, 00:08:33.782 "data_offset": 0, 00:08:33.782 "data_size": 63488 00:08:33.782 }, 00:08:33.782 { 00:08:33.782 "name": "BaseBdev2", 00:08:33.782 "uuid": "0f2b4ad9-50dc-44fc-b2fe-1b3eb7bbeff9", 00:08:33.782 "is_configured": true, 00:08:33.782 "data_offset": 2048, 00:08:33.782 "data_size": 63488 00:08:33.782 }, 00:08:33.782 { 00:08:33.782 "name": "BaseBdev3", 00:08:33.782 "uuid": "2c8a4419-577a-46c7-95e6-3586e99e7865", 00:08:33.782 "is_configured": true, 00:08:33.782 "data_offset": 2048, 00:08:33.782 "data_size": 63488 00:08:33.782 } 00:08:33.782 ] 00:08:33.782 }' 00:08:33.782 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.782 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:34.043 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.044 [2024-09-29 16:46:55.639230] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.044 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.044 [2024-09-29 16:46:55.710194] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:34.044 [2024-09-29 16:46:55.710239] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:34.305 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.305 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:34.305 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:34.305 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 BaseBdev2 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 [ 00:08:34.306 { 00:08:34.306 "name": "BaseBdev2", 00:08:34.306 "aliases": [ 00:08:34.306 "c74939c4-e441-4dcc-95ae-aae47eff8aac" 00:08:34.306 ], 00:08:34.306 "product_name": "Malloc disk", 00:08:34.306 "block_size": 512, 00:08:34.306 "num_blocks": 65536, 00:08:34.306 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:34.306 "assigned_rate_limits": { 00:08:34.306 "rw_ios_per_sec": 0, 00:08:34.306 "rw_mbytes_per_sec": 0, 00:08:34.306 "r_mbytes_per_sec": 0, 00:08:34.306 "w_mbytes_per_sec": 0 00:08:34.306 }, 00:08:34.306 "claimed": false, 00:08:34.306 "zoned": false, 00:08:34.306 "supported_io_types": { 00:08:34.306 "read": true, 00:08:34.306 "write": true, 00:08:34.306 "unmap": true, 00:08:34.306 "flush": true, 00:08:34.306 "reset": true, 00:08:34.306 "nvme_admin": false, 00:08:34.306 "nvme_io": false, 00:08:34.306 "nvme_io_md": false, 00:08:34.306 "write_zeroes": true, 00:08:34.306 "zcopy": true, 00:08:34.306 "get_zone_info": false, 00:08:34.306 "zone_management": false, 00:08:34.306 "zone_append": false, 00:08:34.306 "compare": false, 00:08:34.306 "compare_and_write": false, 00:08:34.306 "abort": true, 00:08:34.306 "seek_hole": false, 00:08:34.306 "seek_data": false, 00:08:34.306 "copy": true, 00:08:34.306 "nvme_iov_md": false 00:08:34.306 }, 00:08:34.306 "memory_domains": [ 00:08:34.306 { 00:08:34.306 "dma_device_id": "system", 00:08:34.306 "dma_device_type": 1 00:08:34.306 }, 00:08:34.306 { 00:08:34.306 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.306 "dma_device_type": 2 00:08:34.306 } 00:08:34.306 ], 00:08:34.306 "driver_specific": {} 00:08:34.306 } 00:08:34.306 ] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 BaseBdev3 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 [ 00:08:34.306 { 00:08:34.306 "name": "BaseBdev3", 00:08:34.306 "aliases": [ 00:08:34.306 "90f7d45d-1a49-4b7d-8176-7d8379b4d79d" 00:08:34.306 ], 00:08:34.306 "product_name": "Malloc disk", 00:08:34.306 "block_size": 512, 00:08:34.306 "num_blocks": 65536, 00:08:34.306 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:34.306 "assigned_rate_limits": { 00:08:34.306 "rw_ios_per_sec": 0, 00:08:34.306 "rw_mbytes_per_sec": 0, 00:08:34.306 "r_mbytes_per_sec": 0, 00:08:34.306 "w_mbytes_per_sec": 0 00:08:34.306 }, 00:08:34.306 "claimed": false, 00:08:34.306 "zoned": false, 00:08:34.306 "supported_io_types": { 00:08:34.306 "read": true, 00:08:34.306 "write": true, 00:08:34.306 "unmap": true, 00:08:34.306 "flush": true, 00:08:34.306 "reset": true, 00:08:34.306 "nvme_admin": false, 00:08:34.306 "nvme_io": false, 00:08:34.306 "nvme_io_md": false, 00:08:34.306 "write_zeroes": true, 00:08:34.306 "zcopy": true, 00:08:34.306 "get_zone_info": false, 00:08:34.306 "zone_management": false, 00:08:34.306 "zone_append": false, 00:08:34.306 "compare": false, 00:08:34.306 "compare_and_write": false, 00:08:34.306 "abort": true, 00:08:34.306 "seek_hole": false, 00:08:34.306 "seek_data": false, 00:08:34.306 "copy": true, 00:08:34.306 "nvme_iov_md": false 00:08:34.306 }, 00:08:34.306 "memory_domains": [ 00:08:34.306 { 00:08:34.306 "dma_device_id": "system", 00:08:34.306 "dma_device_type": 1 00:08:34.306 }, 00:08:34.306 { 00:08:34.306 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.306 "dma_device_type": 2 00:08:34.306 } 00:08:34.306 ], 00:08:34.306 "driver_specific": {} 00:08:34.306 } 00:08:34.306 ] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.306 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.306 [2024-09-29 16:46:55.885151] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:34.306 [2024-09-29 16:46:55.885247] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:34.306 [2024-09-29 16:46:55.885286] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:34.307 [2024-09-29 16:46:55.887111] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.307 "name": "Existed_Raid", 00:08:34.307 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:34.307 "strip_size_kb": 64, 00:08:34.307 "state": "configuring", 00:08:34.307 "raid_level": "concat", 00:08:34.307 "superblock": true, 00:08:34.307 "num_base_bdevs": 3, 00:08:34.307 "num_base_bdevs_discovered": 2, 00:08:34.307 "num_base_bdevs_operational": 3, 00:08:34.307 "base_bdevs_list": [ 00:08:34.307 { 00:08:34.307 "name": "BaseBdev1", 00:08:34.307 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:34.307 "is_configured": false, 00:08:34.307 "data_offset": 0, 00:08:34.307 "data_size": 0 00:08:34.307 }, 00:08:34.307 { 00:08:34.307 "name": "BaseBdev2", 00:08:34.307 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:34.307 "is_configured": true, 00:08:34.307 "data_offset": 2048, 00:08:34.307 "data_size": 63488 00:08:34.307 }, 00:08:34.307 { 00:08:34.307 "name": "BaseBdev3", 00:08:34.307 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:34.307 "is_configured": true, 00:08:34.307 "data_offset": 2048, 00:08:34.307 "data_size": 63488 00:08:34.307 } 00:08:34.307 ] 00:08:34.307 }' 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.307 16:46:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.876 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:34.876 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.876 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.876 [2024-09-29 16:46:56.372306] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:34.876 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:34.877 "name": "Existed_Raid", 00:08:34.877 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:34.877 "strip_size_kb": 64, 00:08:34.877 "state": "configuring", 00:08:34.877 "raid_level": "concat", 00:08:34.877 "superblock": true, 00:08:34.877 "num_base_bdevs": 3, 00:08:34.877 "num_base_bdevs_discovered": 1, 00:08:34.877 "num_base_bdevs_operational": 3, 00:08:34.877 "base_bdevs_list": [ 00:08:34.877 { 00:08:34.877 "name": "BaseBdev1", 00:08:34.877 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:34.877 "is_configured": false, 00:08:34.877 "data_offset": 0, 00:08:34.877 "data_size": 0 00:08:34.877 }, 00:08:34.877 { 00:08:34.877 "name": null, 00:08:34.877 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:34.877 "is_configured": false, 00:08:34.877 "data_offset": 0, 00:08:34.877 "data_size": 63488 00:08:34.877 }, 00:08:34.877 { 00:08:34.877 "name": "BaseBdev3", 00:08:34.877 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:34.877 "is_configured": true, 00:08:34.877 "data_offset": 2048, 00:08:34.877 "data_size": 63488 00:08:34.877 } 00:08:34.877 ] 00:08:34.877 }' 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:34.877 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.447 [2024-09-29 16:46:56.882363] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:35.447 BaseBdev1 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.447 [ 00:08:35.447 { 00:08:35.447 "name": "BaseBdev1", 00:08:35.447 "aliases": [ 00:08:35.447 "d63967bc-5a53-47eb-a156-2ece452f48a6" 00:08:35.447 ], 00:08:35.447 "product_name": "Malloc disk", 00:08:35.447 "block_size": 512, 00:08:35.447 "num_blocks": 65536, 00:08:35.447 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:35.447 "assigned_rate_limits": { 00:08:35.447 "rw_ios_per_sec": 0, 00:08:35.447 "rw_mbytes_per_sec": 0, 00:08:35.447 "r_mbytes_per_sec": 0, 00:08:35.447 "w_mbytes_per_sec": 0 00:08:35.447 }, 00:08:35.447 "claimed": true, 00:08:35.447 "claim_type": "exclusive_write", 00:08:35.447 "zoned": false, 00:08:35.447 "supported_io_types": { 00:08:35.447 "read": true, 00:08:35.447 "write": true, 00:08:35.447 "unmap": true, 00:08:35.447 "flush": true, 00:08:35.447 "reset": true, 00:08:35.447 "nvme_admin": false, 00:08:35.447 "nvme_io": false, 00:08:35.447 "nvme_io_md": false, 00:08:35.447 "write_zeroes": true, 00:08:35.447 "zcopy": true, 00:08:35.447 "get_zone_info": false, 00:08:35.447 "zone_management": false, 00:08:35.447 "zone_append": false, 00:08:35.447 "compare": false, 00:08:35.447 "compare_and_write": false, 00:08:35.447 "abort": true, 00:08:35.447 "seek_hole": false, 00:08:35.447 "seek_data": false, 00:08:35.447 "copy": true, 00:08:35.447 "nvme_iov_md": false 00:08:35.447 }, 00:08:35.447 "memory_domains": [ 00:08:35.447 { 00:08:35.447 "dma_device_id": "system", 00:08:35.447 "dma_device_type": 1 00:08:35.447 }, 00:08:35.447 { 00:08:35.447 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:35.447 "dma_device_type": 2 00:08:35.447 } 00:08:35.447 ], 00:08:35.447 "driver_specific": {} 00:08:35.447 } 00:08:35.447 ] 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.447 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.448 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:35.448 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.448 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.448 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.448 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.448 "name": "Existed_Raid", 00:08:35.448 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:35.448 "strip_size_kb": 64, 00:08:35.448 "state": "configuring", 00:08:35.448 "raid_level": "concat", 00:08:35.448 "superblock": true, 00:08:35.448 "num_base_bdevs": 3, 00:08:35.448 "num_base_bdevs_discovered": 2, 00:08:35.448 "num_base_bdevs_operational": 3, 00:08:35.448 "base_bdevs_list": [ 00:08:35.448 { 00:08:35.448 "name": "BaseBdev1", 00:08:35.448 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:35.448 "is_configured": true, 00:08:35.448 "data_offset": 2048, 00:08:35.448 "data_size": 63488 00:08:35.448 }, 00:08:35.448 { 00:08:35.448 "name": null, 00:08:35.448 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:35.448 "is_configured": false, 00:08:35.448 "data_offset": 0, 00:08:35.448 "data_size": 63488 00:08:35.448 }, 00:08:35.448 { 00:08:35.448 "name": "BaseBdev3", 00:08:35.448 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:35.448 "is_configured": true, 00:08:35.448 "data_offset": 2048, 00:08:35.448 "data_size": 63488 00:08:35.448 } 00:08:35.448 ] 00:08:35.448 }' 00:08:35.448 16:46:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.448 16:46:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.707 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.707 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:35.707 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.708 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.708 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.967 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:35.967 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:35.967 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.967 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.967 [2024-09-29 16:46:57.397601] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.968 "name": "Existed_Raid", 00:08:35.968 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:35.968 "strip_size_kb": 64, 00:08:35.968 "state": "configuring", 00:08:35.968 "raid_level": "concat", 00:08:35.968 "superblock": true, 00:08:35.968 "num_base_bdevs": 3, 00:08:35.968 "num_base_bdevs_discovered": 1, 00:08:35.968 "num_base_bdevs_operational": 3, 00:08:35.968 "base_bdevs_list": [ 00:08:35.968 { 00:08:35.968 "name": "BaseBdev1", 00:08:35.968 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:35.968 "is_configured": true, 00:08:35.968 "data_offset": 2048, 00:08:35.968 "data_size": 63488 00:08:35.968 }, 00:08:35.968 { 00:08:35.968 "name": null, 00:08:35.968 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:35.968 "is_configured": false, 00:08:35.968 "data_offset": 0, 00:08:35.968 "data_size": 63488 00:08:35.968 }, 00:08:35.968 { 00:08:35.968 "name": null, 00:08:35.968 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:35.968 "is_configured": false, 00:08:35.968 "data_offset": 0, 00:08:35.968 "data_size": 63488 00:08:35.968 } 00:08:35.968 ] 00:08:35.968 }' 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.968 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.228 [2024-09-29 16:46:57.852840] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.228 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.488 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.488 "name": "Existed_Raid", 00:08:36.488 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:36.488 "strip_size_kb": 64, 00:08:36.488 "state": "configuring", 00:08:36.488 "raid_level": "concat", 00:08:36.488 "superblock": true, 00:08:36.488 "num_base_bdevs": 3, 00:08:36.488 "num_base_bdevs_discovered": 2, 00:08:36.488 "num_base_bdevs_operational": 3, 00:08:36.488 "base_bdevs_list": [ 00:08:36.488 { 00:08:36.488 "name": "BaseBdev1", 00:08:36.488 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:36.488 "is_configured": true, 00:08:36.488 "data_offset": 2048, 00:08:36.488 "data_size": 63488 00:08:36.488 }, 00:08:36.488 { 00:08:36.488 "name": null, 00:08:36.488 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:36.488 "is_configured": false, 00:08:36.488 "data_offset": 0, 00:08:36.488 "data_size": 63488 00:08:36.488 }, 00:08:36.488 { 00:08:36.488 "name": "BaseBdev3", 00:08:36.488 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:36.488 "is_configured": true, 00:08:36.488 "data_offset": 2048, 00:08:36.488 "data_size": 63488 00:08:36.488 } 00:08:36.488 ] 00:08:36.488 }' 00:08:36.488 16:46:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.488 16:46:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.748 [2024-09-29 16:46:58.371992] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.748 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.008 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.008 "name": "Existed_Raid", 00:08:37.008 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:37.008 "strip_size_kb": 64, 00:08:37.008 "state": "configuring", 00:08:37.008 "raid_level": "concat", 00:08:37.008 "superblock": true, 00:08:37.008 "num_base_bdevs": 3, 00:08:37.008 "num_base_bdevs_discovered": 1, 00:08:37.008 "num_base_bdevs_operational": 3, 00:08:37.008 "base_bdevs_list": [ 00:08:37.008 { 00:08:37.008 "name": null, 00:08:37.008 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:37.008 "is_configured": false, 00:08:37.008 "data_offset": 0, 00:08:37.008 "data_size": 63488 00:08:37.008 }, 00:08:37.008 { 00:08:37.008 "name": null, 00:08:37.008 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:37.008 "is_configured": false, 00:08:37.008 "data_offset": 0, 00:08:37.008 "data_size": 63488 00:08:37.008 }, 00:08:37.008 { 00:08:37.008 "name": "BaseBdev3", 00:08:37.008 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:37.008 "is_configured": true, 00:08:37.008 "data_offset": 2048, 00:08:37.008 "data_size": 63488 00:08:37.008 } 00:08:37.008 ] 00:08:37.008 }' 00:08:37.008 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.008 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.267 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.267 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.268 [2024-09-29 16:46:58.893492] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.268 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.527 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.527 "name": "Existed_Raid", 00:08:37.527 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:37.527 "strip_size_kb": 64, 00:08:37.527 "state": "configuring", 00:08:37.527 "raid_level": "concat", 00:08:37.527 "superblock": true, 00:08:37.527 "num_base_bdevs": 3, 00:08:37.527 "num_base_bdevs_discovered": 2, 00:08:37.527 "num_base_bdevs_operational": 3, 00:08:37.527 "base_bdevs_list": [ 00:08:37.527 { 00:08:37.527 "name": null, 00:08:37.527 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:37.527 "is_configured": false, 00:08:37.527 "data_offset": 0, 00:08:37.527 "data_size": 63488 00:08:37.527 }, 00:08:37.527 { 00:08:37.527 "name": "BaseBdev2", 00:08:37.527 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:37.527 "is_configured": true, 00:08:37.527 "data_offset": 2048, 00:08:37.527 "data_size": 63488 00:08:37.527 }, 00:08:37.527 { 00:08:37.527 "name": "BaseBdev3", 00:08:37.527 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:37.527 "is_configured": true, 00:08:37.527 "data_offset": 2048, 00:08:37.527 "data_size": 63488 00:08:37.527 } 00:08:37.527 ] 00:08:37.527 }' 00:08:37.527 16:46:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.527 16:46:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d63967bc-5a53-47eb-a156-2ece452f48a6 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.785 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.045 NewBaseBdev 00:08:38.045 [2024-09-29 16:46:59.467485] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:38.045 [2024-09-29 16:46:59.467651] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:38.045 [2024-09-29 16:46:59.467667] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:38.045 [2024-09-29 16:46:59.467939] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:38.045 [2024-09-29 16:46:59.468054] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:38.045 [2024-09-29 16:46:59.468063] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:38.045 [2024-09-29 16:46:59.468169] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.045 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.045 [ 00:08:38.045 { 00:08:38.045 "name": "NewBaseBdev", 00:08:38.045 "aliases": [ 00:08:38.045 "d63967bc-5a53-47eb-a156-2ece452f48a6" 00:08:38.045 ], 00:08:38.045 "product_name": "Malloc disk", 00:08:38.045 "block_size": 512, 00:08:38.045 "num_blocks": 65536, 00:08:38.045 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:38.045 "assigned_rate_limits": { 00:08:38.045 "rw_ios_per_sec": 0, 00:08:38.045 "rw_mbytes_per_sec": 0, 00:08:38.045 "r_mbytes_per_sec": 0, 00:08:38.045 "w_mbytes_per_sec": 0 00:08:38.045 }, 00:08:38.045 "claimed": true, 00:08:38.045 "claim_type": "exclusive_write", 00:08:38.045 "zoned": false, 00:08:38.045 "supported_io_types": { 00:08:38.046 "read": true, 00:08:38.046 "write": true, 00:08:38.046 "unmap": true, 00:08:38.046 "flush": true, 00:08:38.046 "reset": true, 00:08:38.046 "nvme_admin": false, 00:08:38.046 "nvme_io": false, 00:08:38.046 "nvme_io_md": false, 00:08:38.046 "write_zeroes": true, 00:08:38.046 "zcopy": true, 00:08:38.046 "get_zone_info": false, 00:08:38.046 "zone_management": false, 00:08:38.046 "zone_append": false, 00:08:38.046 "compare": false, 00:08:38.046 "compare_and_write": false, 00:08:38.046 "abort": true, 00:08:38.046 "seek_hole": false, 00:08:38.046 "seek_data": false, 00:08:38.046 "copy": true, 00:08:38.046 "nvme_iov_md": false 00:08:38.046 }, 00:08:38.046 "memory_domains": [ 00:08:38.046 { 00:08:38.046 "dma_device_id": "system", 00:08:38.046 "dma_device_type": 1 00:08:38.046 }, 00:08:38.046 { 00:08:38.046 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.046 "dma_device_type": 2 00:08:38.046 } 00:08:38.046 ], 00:08:38.046 "driver_specific": {} 00:08:38.046 } 00:08:38.046 ] 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.046 "name": "Existed_Raid", 00:08:38.046 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:38.046 "strip_size_kb": 64, 00:08:38.046 "state": "online", 00:08:38.046 "raid_level": "concat", 00:08:38.046 "superblock": true, 00:08:38.046 "num_base_bdevs": 3, 00:08:38.046 "num_base_bdevs_discovered": 3, 00:08:38.046 "num_base_bdevs_operational": 3, 00:08:38.046 "base_bdevs_list": [ 00:08:38.046 { 00:08:38.046 "name": "NewBaseBdev", 00:08:38.046 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:38.046 "is_configured": true, 00:08:38.046 "data_offset": 2048, 00:08:38.046 "data_size": 63488 00:08:38.046 }, 00:08:38.046 { 00:08:38.046 "name": "BaseBdev2", 00:08:38.046 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:38.046 "is_configured": true, 00:08:38.046 "data_offset": 2048, 00:08:38.046 "data_size": 63488 00:08:38.046 }, 00:08:38.046 { 00:08:38.046 "name": "BaseBdev3", 00:08:38.046 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:38.046 "is_configured": true, 00:08:38.046 "data_offset": 2048, 00:08:38.046 "data_size": 63488 00:08:38.046 } 00:08:38.046 ] 00:08:38.046 }' 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.046 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:38.306 [2024-09-29 16:46:59.958977] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:38.306 16:46:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.566 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:38.566 "name": "Existed_Raid", 00:08:38.566 "aliases": [ 00:08:38.566 "de020cf2-3d3c-4b5f-940e-2b1338e67247" 00:08:38.566 ], 00:08:38.566 "product_name": "Raid Volume", 00:08:38.566 "block_size": 512, 00:08:38.566 "num_blocks": 190464, 00:08:38.566 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:38.566 "assigned_rate_limits": { 00:08:38.566 "rw_ios_per_sec": 0, 00:08:38.566 "rw_mbytes_per_sec": 0, 00:08:38.566 "r_mbytes_per_sec": 0, 00:08:38.566 "w_mbytes_per_sec": 0 00:08:38.566 }, 00:08:38.566 "claimed": false, 00:08:38.566 "zoned": false, 00:08:38.566 "supported_io_types": { 00:08:38.566 "read": true, 00:08:38.566 "write": true, 00:08:38.566 "unmap": true, 00:08:38.566 "flush": true, 00:08:38.566 "reset": true, 00:08:38.566 "nvme_admin": false, 00:08:38.566 "nvme_io": false, 00:08:38.566 "nvme_io_md": false, 00:08:38.566 "write_zeroes": true, 00:08:38.566 "zcopy": false, 00:08:38.566 "get_zone_info": false, 00:08:38.566 "zone_management": false, 00:08:38.566 "zone_append": false, 00:08:38.566 "compare": false, 00:08:38.566 "compare_and_write": false, 00:08:38.566 "abort": false, 00:08:38.566 "seek_hole": false, 00:08:38.566 "seek_data": false, 00:08:38.566 "copy": false, 00:08:38.566 "nvme_iov_md": false 00:08:38.566 }, 00:08:38.566 "memory_domains": [ 00:08:38.566 { 00:08:38.566 "dma_device_id": "system", 00:08:38.566 "dma_device_type": 1 00:08:38.566 }, 00:08:38.566 { 00:08:38.566 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.566 "dma_device_type": 2 00:08:38.566 }, 00:08:38.566 { 00:08:38.566 "dma_device_id": "system", 00:08:38.566 "dma_device_type": 1 00:08:38.566 }, 00:08:38.566 { 00:08:38.566 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.566 "dma_device_type": 2 00:08:38.566 }, 00:08:38.566 { 00:08:38.566 "dma_device_id": "system", 00:08:38.566 "dma_device_type": 1 00:08:38.566 }, 00:08:38.566 { 00:08:38.566 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.566 "dma_device_type": 2 00:08:38.566 } 00:08:38.566 ], 00:08:38.566 "driver_specific": { 00:08:38.566 "raid": { 00:08:38.566 "uuid": "de020cf2-3d3c-4b5f-940e-2b1338e67247", 00:08:38.566 "strip_size_kb": 64, 00:08:38.566 "state": "online", 00:08:38.566 "raid_level": "concat", 00:08:38.566 "superblock": true, 00:08:38.566 "num_base_bdevs": 3, 00:08:38.566 "num_base_bdevs_discovered": 3, 00:08:38.566 "num_base_bdevs_operational": 3, 00:08:38.566 "base_bdevs_list": [ 00:08:38.566 { 00:08:38.566 "name": "NewBaseBdev", 00:08:38.566 "uuid": "d63967bc-5a53-47eb-a156-2ece452f48a6", 00:08:38.566 "is_configured": true, 00:08:38.566 "data_offset": 2048, 00:08:38.566 "data_size": 63488 00:08:38.566 }, 00:08:38.566 { 00:08:38.566 "name": "BaseBdev2", 00:08:38.566 "uuid": "c74939c4-e441-4dcc-95ae-aae47eff8aac", 00:08:38.566 "is_configured": true, 00:08:38.566 "data_offset": 2048, 00:08:38.566 "data_size": 63488 00:08:38.566 }, 00:08:38.566 { 00:08:38.566 "name": "BaseBdev3", 00:08:38.566 "uuid": "90f7d45d-1a49-4b7d-8176-7d8379b4d79d", 00:08:38.566 "is_configured": true, 00:08:38.566 "data_offset": 2048, 00:08:38.566 "data_size": 63488 00:08:38.566 } 00:08:38.566 ] 00:08:38.566 } 00:08:38.566 } 00:08:38.566 }' 00:08:38.566 16:46:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:38.566 BaseBdev2 00:08:38.566 BaseBdev3' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.566 [2024-09-29 16:47:00.226206] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:38.566 [2024-09-29 16:47:00.226269] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:38.566 [2024-09-29 16:47:00.226357] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:38.566 [2024-09-29 16:47:00.226432] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:38.566 [2024-09-29 16:47:00.226477] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.566 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 77051 00:08:38.567 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 77051 ']' 00:08:38.567 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 77051 00:08:38.567 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:08:38.826 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:38.826 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77051 00:08:38.826 killing process with pid 77051 00:08:38.826 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:38.826 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:38.826 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77051' 00:08:38.826 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 77051 00:08:38.826 [2024-09-29 16:47:00.275229] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:38.826 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 77051 00:08:38.826 [2024-09-29 16:47:00.305810] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:39.088 16:47:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:39.088 00:08:39.088 real 0m9.014s 00:08:39.088 user 0m15.450s 00:08:39.088 sys 0m1.755s 00:08:39.088 ************************************ 00:08:39.088 END TEST raid_state_function_test_sb 00:08:39.088 ************************************ 00:08:39.088 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:39.088 16:47:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.088 16:47:00 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:08:39.088 16:47:00 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:39.088 16:47:00 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:39.088 16:47:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:39.088 ************************************ 00:08:39.088 START TEST raid_superblock_test 00:08:39.088 ************************************ 00:08:39.088 16:47:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 3 00:08:39.088 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:08:39.088 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:39.088 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=77654 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 77654 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 77654 ']' 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:39.089 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:39.089 16:47:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.089 [2024-09-29 16:47:00.704699] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:39.089 [2024-09-29 16:47:00.705339] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77654 ] 00:08:39.348 [2024-09-29 16:47:00.850246] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:39.348 [2024-09-29 16:47:00.896182] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:39.348 [2024-09-29 16:47:00.938212] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.348 [2024-09-29 16:47:00.938247] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.918 malloc1 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.918 [2024-09-29 16:47:01.540112] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:39.918 [2024-09-29 16:47:01.540223] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:39.918 [2024-09-29 16:47:01.540261] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:39.918 [2024-09-29 16:47:01.540294] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:39.918 [2024-09-29 16:47:01.542422] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:39.918 [2024-09-29 16:47:01.542494] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:39.918 pt1 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.918 malloc2 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:39.918 [2024-09-29 16:47:01.581386] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:39.918 [2024-09-29 16:47:01.581495] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:39.918 [2024-09-29 16:47:01.581535] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:39.918 [2024-09-29 16:47:01.581581] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:39.918 [2024-09-29 16:47:01.583810] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:39.918 [2024-09-29 16:47:01.583881] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:39.918 pt2 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.918 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.178 malloc3 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.178 [2024-09-29 16:47:01.609930] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:40.178 [2024-09-29 16:47:01.610019] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:40.178 [2024-09-29 16:47:01.610067] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:40.178 [2024-09-29 16:47:01.610097] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:40.178 [2024-09-29 16:47:01.612249] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:40.178 [2024-09-29 16:47:01.612322] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:40.178 pt3 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.178 [2024-09-29 16:47:01.622001] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:40.178 [2024-09-29 16:47:01.623877] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:40.178 [2024-09-29 16:47:01.623973] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:40.178 [2024-09-29 16:47:01.624172] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:40.178 [2024-09-29 16:47:01.624237] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:40.178 [2024-09-29 16:47:01.624574] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:40.178 [2024-09-29 16:47:01.624772] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:40.178 [2024-09-29 16:47:01.624822] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:40.178 [2024-09-29 16:47:01.625021] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:40.178 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.179 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.179 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.179 "name": "raid_bdev1", 00:08:40.179 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:40.179 "strip_size_kb": 64, 00:08:40.179 "state": "online", 00:08:40.179 "raid_level": "concat", 00:08:40.179 "superblock": true, 00:08:40.179 "num_base_bdevs": 3, 00:08:40.179 "num_base_bdevs_discovered": 3, 00:08:40.179 "num_base_bdevs_operational": 3, 00:08:40.179 "base_bdevs_list": [ 00:08:40.179 { 00:08:40.179 "name": "pt1", 00:08:40.179 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:40.179 "is_configured": true, 00:08:40.179 "data_offset": 2048, 00:08:40.179 "data_size": 63488 00:08:40.179 }, 00:08:40.179 { 00:08:40.179 "name": "pt2", 00:08:40.179 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:40.179 "is_configured": true, 00:08:40.179 "data_offset": 2048, 00:08:40.179 "data_size": 63488 00:08:40.179 }, 00:08:40.179 { 00:08:40.179 "name": "pt3", 00:08:40.179 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:40.179 "is_configured": true, 00:08:40.179 "data_offset": 2048, 00:08:40.179 "data_size": 63488 00:08:40.179 } 00:08:40.179 ] 00:08:40.179 }' 00:08:40.179 16:47:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.179 16:47:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.437 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.437 [2024-09-29 16:47:02.101403] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:40.698 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.698 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:40.698 "name": "raid_bdev1", 00:08:40.698 "aliases": [ 00:08:40.698 "52853b09-4ee1-4d75-8025-830ecc880692" 00:08:40.698 ], 00:08:40.698 "product_name": "Raid Volume", 00:08:40.698 "block_size": 512, 00:08:40.698 "num_blocks": 190464, 00:08:40.698 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:40.698 "assigned_rate_limits": { 00:08:40.698 "rw_ios_per_sec": 0, 00:08:40.698 "rw_mbytes_per_sec": 0, 00:08:40.698 "r_mbytes_per_sec": 0, 00:08:40.698 "w_mbytes_per_sec": 0 00:08:40.698 }, 00:08:40.698 "claimed": false, 00:08:40.698 "zoned": false, 00:08:40.698 "supported_io_types": { 00:08:40.698 "read": true, 00:08:40.698 "write": true, 00:08:40.698 "unmap": true, 00:08:40.698 "flush": true, 00:08:40.698 "reset": true, 00:08:40.698 "nvme_admin": false, 00:08:40.698 "nvme_io": false, 00:08:40.698 "nvme_io_md": false, 00:08:40.698 "write_zeroes": true, 00:08:40.698 "zcopy": false, 00:08:40.698 "get_zone_info": false, 00:08:40.698 "zone_management": false, 00:08:40.698 "zone_append": false, 00:08:40.698 "compare": false, 00:08:40.698 "compare_and_write": false, 00:08:40.698 "abort": false, 00:08:40.698 "seek_hole": false, 00:08:40.698 "seek_data": false, 00:08:40.698 "copy": false, 00:08:40.698 "nvme_iov_md": false 00:08:40.698 }, 00:08:40.698 "memory_domains": [ 00:08:40.698 { 00:08:40.698 "dma_device_id": "system", 00:08:40.698 "dma_device_type": 1 00:08:40.698 }, 00:08:40.698 { 00:08:40.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.698 "dma_device_type": 2 00:08:40.698 }, 00:08:40.698 { 00:08:40.698 "dma_device_id": "system", 00:08:40.698 "dma_device_type": 1 00:08:40.698 }, 00:08:40.698 { 00:08:40.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.698 "dma_device_type": 2 00:08:40.698 }, 00:08:40.698 { 00:08:40.698 "dma_device_id": "system", 00:08:40.698 "dma_device_type": 1 00:08:40.698 }, 00:08:40.698 { 00:08:40.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:40.698 "dma_device_type": 2 00:08:40.698 } 00:08:40.698 ], 00:08:40.698 "driver_specific": { 00:08:40.698 "raid": { 00:08:40.698 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:40.698 "strip_size_kb": 64, 00:08:40.698 "state": "online", 00:08:40.698 "raid_level": "concat", 00:08:40.699 "superblock": true, 00:08:40.699 "num_base_bdevs": 3, 00:08:40.699 "num_base_bdevs_discovered": 3, 00:08:40.699 "num_base_bdevs_operational": 3, 00:08:40.699 "base_bdevs_list": [ 00:08:40.699 { 00:08:40.699 "name": "pt1", 00:08:40.699 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:40.699 "is_configured": true, 00:08:40.699 "data_offset": 2048, 00:08:40.699 "data_size": 63488 00:08:40.699 }, 00:08:40.699 { 00:08:40.699 "name": "pt2", 00:08:40.699 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:40.699 "is_configured": true, 00:08:40.699 "data_offset": 2048, 00:08:40.699 "data_size": 63488 00:08:40.699 }, 00:08:40.699 { 00:08:40.699 "name": "pt3", 00:08:40.699 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:40.699 "is_configured": true, 00:08:40.699 "data_offset": 2048, 00:08:40.699 "data_size": 63488 00:08:40.699 } 00:08:40.699 ] 00:08:40.699 } 00:08:40.699 } 00:08:40.699 }' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:40.699 pt2 00:08:40.699 pt3' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:40.699 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 [2024-09-29 16:47:02.400859] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=52853b09-4ee1-4d75-8025-830ecc880692 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 52853b09-4ee1-4d75-8025-830ecc880692 ']' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 [2024-09-29 16:47:02.448495] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:40.959 [2024-09-29 16:47:02.448523] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:40.959 [2024-09-29 16:47:02.448595] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:40.959 [2024-09-29 16:47:02.448652] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:40.959 [2024-09-29 16:47:02.448665] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 [2024-09-29 16:47:02.592325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:40.959 [2024-09-29 16:47:02.594138] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:40.959 [2024-09-29 16:47:02.594199] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:40.959 [2024-09-29 16:47:02.594256] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:40.959 [2024-09-29 16:47:02.594297] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:40.959 [2024-09-29 16:47:02.594316] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:40.959 [2024-09-29 16:47:02.594328] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:40.959 [2024-09-29 16:47:02.594338] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:40.959 request: 00:08:40.959 { 00:08:40.959 "name": "raid_bdev1", 00:08:40.959 "raid_level": "concat", 00:08:40.959 "base_bdevs": [ 00:08:40.959 "malloc1", 00:08:40.959 "malloc2", 00:08:40.959 "malloc3" 00:08:40.959 ], 00:08:40.959 "strip_size_kb": 64, 00:08:40.959 "superblock": false, 00:08:40.959 "method": "bdev_raid_create", 00:08:40.959 "req_id": 1 00:08:40.959 } 00:08:40.959 Got JSON-RPC error response 00:08:40.959 response: 00:08:40.959 { 00:08:40.959 "code": -17, 00:08:40.959 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:40.959 } 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:40.959 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.219 [2024-09-29 16:47:02.664177] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:41.219 [2024-09-29 16:47:02.664229] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:41.219 [2024-09-29 16:47:02.664244] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:41.219 [2024-09-29 16:47:02.664254] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:41.219 [2024-09-29 16:47:02.666438] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:41.219 [2024-09-29 16:47:02.666476] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:41.219 [2024-09-29 16:47:02.666557] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:41.219 [2024-09-29 16:47:02.666600] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:41.219 pt1 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.219 "name": "raid_bdev1", 00:08:41.219 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:41.219 "strip_size_kb": 64, 00:08:41.219 "state": "configuring", 00:08:41.219 "raid_level": "concat", 00:08:41.219 "superblock": true, 00:08:41.219 "num_base_bdevs": 3, 00:08:41.219 "num_base_bdevs_discovered": 1, 00:08:41.219 "num_base_bdevs_operational": 3, 00:08:41.219 "base_bdevs_list": [ 00:08:41.219 { 00:08:41.219 "name": "pt1", 00:08:41.219 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:41.219 "is_configured": true, 00:08:41.219 "data_offset": 2048, 00:08:41.219 "data_size": 63488 00:08:41.219 }, 00:08:41.219 { 00:08:41.219 "name": null, 00:08:41.219 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:41.219 "is_configured": false, 00:08:41.219 "data_offset": 2048, 00:08:41.219 "data_size": 63488 00:08:41.219 }, 00:08:41.219 { 00:08:41.219 "name": null, 00:08:41.219 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:41.219 "is_configured": false, 00:08:41.219 "data_offset": 2048, 00:08:41.219 "data_size": 63488 00:08:41.219 } 00:08:41.219 ] 00:08:41.219 }' 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.219 16:47:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.479 [2024-09-29 16:47:03.079471] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:41.479 [2024-09-29 16:47:03.079536] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:41.479 [2024-09-29 16:47:03.079556] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:41.479 [2024-09-29 16:47:03.079569] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:41.479 [2024-09-29 16:47:03.079972] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:41.479 [2024-09-29 16:47:03.080001] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:41.479 [2024-09-29 16:47:03.080071] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:41.479 [2024-09-29 16:47:03.080096] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:41.479 pt2 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.479 [2024-09-29 16:47:03.091465] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.479 "name": "raid_bdev1", 00:08:41.479 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:41.479 "strip_size_kb": 64, 00:08:41.479 "state": "configuring", 00:08:41.479 "raid_level": "concat", 00:08:41.479 "superblock": true, 00:08:41.479 "num_base_bdevs": 3, 00:08:41.479 "num_base_bdevs_discovered": 1, 00:08:41.479 "num_base_bdevs_operational": 3, 00:08:41.479 "base_bdevs_list": [ 00:08:41.479 { 00:08:41.479 "name": "pt1", 00:08:41.479 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:41.479 "is_configured": true, 00:08:41.479 "data_offset": 2048, 00:08:41.479 "data_size": 63488 00:08:41.479 }, 00:08:41.479 { 00:08:41.479 "name": null, 00:08:41.479 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:41.479 "is_configured": false, 00:08:41.479 "data_offset": 0, 00:08:41.479 "data_size": 63488 00:08:41.479 }, 00:08:41.479 { 00:08:41.479 "name": null, 00:08:41.479 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:41.479 "is_configured": false, 00:08:41.479 "data_offset": 2048, 00:08:41.479 "data_size": 63488 00:08:41.479 } 00:08:41.479 ] 00:08:41.479 }' 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.479 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.051 [2024-09-29 16:47:03.522731] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:42.051 [2024-09-29 16:47:03.522793] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:42.051 [2024-09-29 16:47:03.522810] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:42.051 [2024-09-29 16:47:03.522818] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:42.051 [2024-09-29 16:47:03.523189] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:42.051 [2024-09-29 16:47:03.523213] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:42.051 [2024-09-29 16:47:03.523274] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:42.051 [2024-09-29 16:47:03.523292] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:42.051 pt2 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.051 [2024-09-29 16:47:03.534703] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:42.051 [2024-09-29 16:47:03.534770] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:42.051 [2024-09-29 16:47:03.534788] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:42.051 [2024-09-29 16:47:03.534795] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:42.051 [2024-09-29 16:47:03.535108] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:42.051 [2024-09-29 16:47:03.535144] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:42.051 [2024-09-29 16:47:03.535216] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:42.051 [2024-09-29 16:47:03.535244] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:42.051 [2024-09-29 16:47:03.535350] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:42.051 [2024-09-29 16:47:03.535373] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:42.051 [2024-09-29 16:47:03.535604] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:42.051 [2024-09-29 16:47:03.535723] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:42.051 [2024-09-29 16:47:03.535755] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:42.051 [2024-09-29 16:47:03.535856] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:42.051 pt3 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.051 "name": "raid_bdev1", 00:08:42.051 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:42.051 "strip_size_kb": 64, 00:08:42.051 "state": "online", 00:08:42.051 "raid_level": "concat", 00:08:42.051 "superblock": true, 00:08:42.051 "num_base_bdevs": 3, 00:08:42.051 "num_base_bdevs_discovered": 3, 00:08:42.051 "num_base_bdevs_operational": 3, 00:08:42.051 "base_bdevs_list": [ 00:08:42.051 { 00:08:42.051 "name": "pt1", 00:08:42.051 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:42.051 "is_configured": true, 00:08:42.051 "data_offset": 2048, 00:08:42.051 "data_size": 63488 00:08:42.051 }, 00:08:42.051 { 00:08:42.051 "name": "pt2", 00:08:42.051 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:42.051 "is_configured": true, 00:08:42.051 "data_offset": 2048, 00:08:42.051 "data_size": 63488 00:08:42.051 }, 00:08:42.051 { 00:08:42.051 "name": "pt3", 00:08:42.051 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:42.051 "is_configured": true, 00:08:42.051 "data_offset": 2048, 00:08:42.051 "data_size": 63488 00:08:42.051 } 00:08:42.051 ] 00:08:42.051 }' 00:08:42.051 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.052 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.324 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:42.324 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:42.324 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:42.324 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:42.324 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:42.324 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:42.592 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:42.592 16:47:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:42.592 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.592 16:47:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.592 [2024-09-29 16:47:04.002230] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:42.592 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.592 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:42.592 "name": "raid_bdev1", 00:08:42.592 "aliases": [ 00:08:42.592 "52853b09-4ee1-4d75-8025-830ecc880692" 00:08:42.592 ], 00:08:42.592 "product_name": "Raid Volume", 00:08:42.592 "block_size": 512, 00:08:42.592 "num_blocks": 190464, 00:08:42.592 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:42.592 "assigned_rate_limits": { 00:08:42.592 "rw_ios_per_sec": 0, 00:08:42.592 "rw_mbytes_per_sec": 0, 00:08:42.592 "r_mbytes_per_sec": 0, 00:08:42.592 "w_mbytes_per_sec": 0 00:08:42.592 }, 00:08:42.592 "claimed": false, 00:08:42.592 "zoned": false, 00:08:42.592 "supported_io_types": { 00:08:42.592 "read": true, 00:08:42.592 "write": true, 00:08:42.592 "unmap": true, 00:08:42.592 "flush": true, 00:08:42.592 "reset": true, 00:08:42.592 "nvme_admin": false, 00:08:42.592 "nvme_io": false, 00:08:42.592 "nvme_io_md": false, 00:08:42.592 "write_zeroes": true, 00:08:42.592 "zcopy": false, 00:08:42.592 "get_zone_info": false, 00:08:42.592 "zone_management": false, 00:08:42.592 "zone_append": false, 00:08:42.592 "compare": false, 00:08:42.592 "compare_and_write": false, 00:08:42.592 "abort": false, 00:08:42.592 "seek_hole": false, 00:08:42.592 "seek_data": false, 00:08:42.592 "copy": false, 00:08:42.592 "nvme_iov_md": false 00:08:42.592 }, 00:08:42.592 "memory_domains": [ 00:08:42.592 { 00:08:42.592 "dma_device_id": "system", 00:08:42.592 "dma_device_type": 1 00:08:42.592 }, 00:08:42.592 { 00:08:42.592 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.592 "dma_device_type": 2 00:08:42.592 }, 00:08:42.592 { 00:08:42.592 "dma_device_id": "system", 00:08:42.592 "dma_device_type": 1 00:08:42.592 }, 00:08:42.592 { 00:08:42.592 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.592 "dma_device_type": 2 00:08:42.592 }, 00:08:42.592 { 00:08:42.592 "dma_device_id": "system", 00:08:42.592 "dma_device_type": 1 00:08:42.592 }, 00:08:42.592 { 00:08:42.592 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.592 "dma_device_type": 2 00:08:42.592 } 00:08:42.592 ], 00:08:42.592 "driver_specific": { 00:08:42.592 "raid": { 00:08:42.592 "uuid": "52853b09-4ee1-4d75-8025-830ecc880692", 00:08:42.592 "strip_size_kb": 64, 00:08:42.592 "state": "online", 00:08:42.592 "raid_level": "concat", 00:08:42.592 "superblock": true, 00:08:42.592 "num_base_bdevs": 3, 00:08:42.592 "num_base_bdevs_discovered": 3, 00:08:42.592 "num_base_bdevs_operational": 3, 00:08:42.592 "base_bdevs_list": [ 00:08:42.592 { 00:08:42.592 "name": "pt1", 00:08:42.592 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:42.592 "is_configured": true, 00:08:42.592 "data_offset": 2048, 00:08:42.592 "data_size": 63488 00:08:42.592 }, 00:08:42.592 { 00:08:42.592 "name": "pt2", 00:08:42.592 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:42.592 "is_configured": true, 00:08:42.592 "data_offset": 2048, 00:08:42.592 "data_size": 63488 00:08:42.592 }, 00:08:42.592 { 00:08:42.592 "name": "pt3", 00:08:42.592 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:42.592 "is_configured": true, 00:08:42.592 "data_offset": 2048, 00:08:42.592 "data_size": 63488 00:08:42.592 } 00:08:42.592 ] 00:08:42.592 } 00:08:42.592 } 00:08:42.592 }' 00:08:42.592 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:42.592 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:42.592 pt2 00:08:42.592 pt3' 00:08:42.592 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.592 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:42.592 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.593 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:42.593 [2024-09-29 16:47:04.249739] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:42.852 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.852 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 52853b09-4ee1-4d75-8025-830ecc880692 '!=' 52853b09-4ee1-4d75-8025-830ecc880692 ']' 00:08:42.852 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 77654 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 77654 ']' 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 77654 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77654 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:42.853 killing process with pid 77654 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77654' 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 77654 00:08:42.853 [2024-09-29 16:47:04.317816] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:42.853 [2024-09-29 16:47:04.317906] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:42.853 [2024-09-29 16:47:04.317976] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:42.853 [2024-09-29 16:47:04.317988] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:42.853 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 77654 00:08:42.853 [2024-09-29 16:47:04.350749] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:43.113 16:47:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:43.113 00:08:43.113 real 0m3.971s 00:08:43.113 user 0m6.299s 00:08:43.113 sys 0m0.807s 00:08:43.113 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:43.113 16:47:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.113 ************************************ 00:08:43.113 END TEST raid_superblock_test 00:08:43.113 ************************************ 00:08:43.113 16:47:04 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:08:43.113 16:47:04 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:43.113 16:47:04 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:43.113 16:47:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:43.113 ************************************ 00:08:43.113 START TEST raid_read_error_test 00:08:43.113 ************************************ 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 read 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.jEQkRN3xsn 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=77891 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 77891 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 77891 ']' 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:43.113 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:43.113 16:47:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.113 [2024-09-29 16:47:04.758892] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:43.113 [2024-09-29 16:47:04.759027] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77891 ] 00:08:43.373 [2024-09-29 16:47:04.903562] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:43.373 [2024-09-29 16:47:04.949086] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:43.373 [2024-09-29 16:47:04.990856] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:43.373 [2024-09-29 16:47:04.990894] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.942 BaseBdev1_malloc 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.942 true 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:43.942 [2024-09-29 16:47:05.604743] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:43.942 [2024-09-29 16:47:05.604803] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:43.942 [2024-09-29 16:47:05.604839] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:43.942 [2024-09-29 16:47:05.604854] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:43.942 [2024-09-29 16:47:05.606946] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:43.942 [2024-09-29 16:47:05.606980] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:43.942 BaseBdev1 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.942 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.203 BaseBdev2_malloc 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.203 true 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.203 [2024-09-29 16:47:05.657174] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:44.203 [2024-09-29 16:47:05.657246] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.203 [2024-09-29 16:47:05.657274] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:44.203 [2024-09-29 16:47:05.657287] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.203 [2024-09-29 16:47:05.660182] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.203 [2024-09-29 16:47:05.660226] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:44.203 BaseBdev2 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.203 BaseBdev3_malloc 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.203 true 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.203 [2024-09-29 16:47:05.697584] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:44.203 [2024-09-29 16:47:05.697628] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.203 [2024-09-29 16:47:05.697645] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:44.203 [2024-09-29 16:47:05.697669] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.203 [2024-09-29 16:47:05.699684] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.203 [2024-09-29 16:47:05.699744] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:44.203 BaseBdev3 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.203 [2024-09-29 16:47:05.709634] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:44.203 [2024-09-29 16:47:05.711473] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:44.203 [2024-09-29 16:47:05.711549] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:44.203 [2024-09-29 16:47:05.711740] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:44.203 [2024-09-29 16:47:05.711763] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:44.203 [2024-09-29 16:47:05.712058] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:44.203 [2024-09-29 16:47:05.712206] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:44.203 [2024-09-29 16:47:05.712224] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:44.203 [2024-09-29 16:47:05.712354] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.203 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:44.204 "name": "raid_bdev1", 00:08:44.204 "uuid": "7a0b0417-030a-42f8-94bd-71fafac0df78", 00:08:44.204 "strip_size_kb": 64, 00:08:44.204 "state": "online", 00:08:44.204 "raid_level": "concat", 00:08:44.204 "superblock": true, 00:08:44.204 "num_base_bdevs": 3, 00:08:44.204 "num_base_bdevs_discovered": 3, 00:08:44.204 "num_base_bdevs_operational": 3, 00:08:44.204 "base_bdevs_list": [ 00:08:44.204 { 00:08:44.204 "name": "BaseBdev1", 00:08:44.204 "uuid": "92550b10-09ca-59c9-ae40-3d80b8b07c90", 00:08:44.204 "is_configured": true, 00:08:44.204 "data_offset": 2048, 00:08:44.204 "data_size": 63488 00:08:44.204 }, 00:08:44.204 { 00:08:44.204 "name": "BaseBdev2", 00:08:44.204 "uuid": "5bf7aaa8-a7b0-50c4-8e5c-5f8952fc7ee2", 00:08:44.204 "is_configured": true, 00:08:44.204 "data_offset": 2048, 00:08:44.204 "data_size": 63488 00:08:44.204 }, 00:08:44.204 { 00:08:44.204 "name": "BaseBdev3", 00:08:44.204 "uuid": "39079e56-d38a-5e95-a9c2-8e38b28ea342", 00:08:44.204 "is_configured": true, 00:08:44.204 "data_offset": 2048, 00:08:44.204 "data_size": 63488 00:08:44.204 } 00:08:44.204 ] 00:08:44.204 }' 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:44.204 16:47:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.772 16:47:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:44.772 16:47:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:44.772 [2024-09-29 16:47:06.269069] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:45.707 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:45.707 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.707 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.707 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.708 "name": "raid_bdev1", 00:08:45.708 "uuid": "7a0b0417-030a-42f8-94bd-71fafac0df78", 00:08:45.708 "strip_size_kb": 64, 00:08:45.708 "state": "online", 00:08:45.708 "raid_level": "concat", 00:08:45.708 "superblock": true, 00:08:45.708 "num_base_bdevs": 3, 00:08:45.708 "num_base_bdevs_discovered": 3, 00:08:45.708 "num_base_bdevs_operational": 3, 00:08:45.708 "base_bdevs_list": [ 00:08:45.708 { 00:08:45.708 "name": "BaseBdev1", 00:08:45.708 "uuid": "92550b10-09ca-59c9-ae40-3d80b8b07c90", 00:08:45.708 "is_configured": true, 00:08:45.708 "data_offset": 2048, 00:08:45.708 "data_size": 63488 00:08:45.708 }, 00:08:45.708 { 00:08:45.708 "name": "BaseBdev2", 00:08:45.708 "uuid": "5bf7aaa8-a7b0-50c4-8e5c-5f8952fc7ee2", 00:08:45.708 "is_configured": true, 00:08:45.708 "data_offset": 2048, 00:08:45.708 "data_size": 63488 00:08:45.708 }, 00:08:45.708 { 00:08:45.708 "name": "BaseBdev3", 00:08:45.708 "uuid": "39079e56-d38a-5e95-a9c2-8e38b28ea342", 00:08:45.708 "is_configured": true, 00:08:45.708 "data_offset": 2048, 00:08:45.708 "data_size": 63488 00:08:45.708 } 00:08:45.708 ] 00:08:45.708 }' 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.708 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.966 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:45.967 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.967 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.967 [2024-09-29 16:47:07.632666] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:45.967 [2024-09-29 16:47:07.632705] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:45.967 [2024-09-29 16:47:07.635178] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:45.967 [2024-09-29 16:47:07.635239] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:45.967 [2024-09-29 16:47:07.635282] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:45.967 [2024-09-29 16:47:07.635298] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:45.967 { 00:08:45.967 "results": [ 00:08:45.967 { 00:08:45.967 "job": "raid_bdev1", 00:08:45.967 "core_mask": "0x1", 00:08:45.967 "workload": "randrw", 00:08:45.967 "percentage": 50, 00:08:45.967 "status": "finished", 00:08:45.967 "queue_depth": 1, 00:08:45.967 "io_size": 131072, 00:08:45.967 "runtime": 1.364396, 00:08:45.967 "iops": 17371.056496794186, 00:08:45.967 "mibps": 2171.3820620992733, 00:08:45.967 "io_failed": 1, 00:08:45.967 "io_timeout": 0, 00:08:45.967 "avg_latency_us": 79.74353462331962, 00:08:45.967 "min_latency_us": 24.593886462882097, 00:08:45.967 "max_latency_us": 1359.3711790393013 00:08:45.967 } 00:08:45.967 ], 00:08:45.967 "core_count": 1 00:08:45.967 } 00:08:45.967 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.967 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 77891 00:08:45.967 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 77891 ']' 00:08:45.967 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 77891 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77891 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:46.225 killing process with pid 77891 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77891' 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 77891 00:08:46.225 [2024-09-29 16:47:07.683235] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:46.225 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 77891 00:08:46.226 [2024-09-29 16:47:07.708279] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.jEQkRN3xsn 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:46.485 00:08:46.485 real 0m3.289s 00:08:46.485 user 0m4.164s 00:08:46.485 sys 0m0.516s 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:46.485 16:47:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.485 ************************************ 00:08:46.485 END TEST raid_read_error_test 00:08:46.485 ************************************ 00:08:46.485 16:47:08 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:08:46.485 16:47:08 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:46.486 16:47:08 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:46.486 16:47:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:46.486 ************************************ 00:08:46.486 START TEST raid_write_error_test 00:08:46.486 ************************************ 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 write 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.C19XyUkWLv 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78031 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78031 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 78031 ']' 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:46.486 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:46.486 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.486 [2024-09-29 16:47:08.122092] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:46.486 [2024-09-29 16:47:08.122218] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78031 ] 00:08:46.746 [2024-09-29 16:47:08.267458] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:46.746 [2024-09-29 16:47:08.313270] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:46.746 [2024-09-29 16:47:08.355112] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:46.746 [2024-09-29 16:47:08.355155] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.315 BaseBdev1_malloc 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.315 true 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.315 [2024-09-29 16:47:08.973148] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:47.315 [2024-09-29 16:47:08.973218] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.315 [2024-09-29 16:47:08.973239] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:47.315 [2024-09-29 16:47:08.973248] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.315 [2024-09-29 16:47:08.975371] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.315 [2024-09-29 16:47:08.975410] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:47.315 BaseBdev1 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.315 16:47:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.574 BaseBdev2_malloc 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.574 true 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.574 [2024-09-29 16:47:09.031154] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:47.574 [2024-09-29 16:47:09.031222] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.574 [2024-09-29 16:47:09.031248] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:47.574 [2024-09-29 16:47:09.031260] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.574 [2024-09-29 16:47:09.033951] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.574 [2024-09-29 16:47:09.033996] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:47.574 BaseBdev2 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:47.574 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.575 BaseBdev3_malloc 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.575 true 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.575 [2024-09-29 16:47:09.071890] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:47.575 [2024-09-29 16:47:09.071934] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:47.575 [2024-09-29 16:47:09.071951] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:47.575 [2024-09-29 16:47:09.071959] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:47.575 [2024-09-29 16:47:09.074073] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:47.575 [2024-09-29 16:47:09.074104] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:47.575 BaseBdev3 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.575 [2024-09-29 16:47:09.083947] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:47.575 [2024-09-29 16:47:09.085802] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:47.575 [2024-09-29 16:47:09.085878] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:47.575 [2024-09-29 16:47:09.086040] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:47.575 [2024-09-29 16:47:09.086078] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:47.575 [2024-09-29 16:47:09.086339] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:47.575 [2024-09-29 16:47:09.086498] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:47.575 [2024-09-29 16:47:09.086516] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:47.575 [2024-09-29 16:47:09.086646] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:47.575 "name": "raid_bdev1", 00:08:47.575 "uuid": "90a7ca6c-1b1a-4047-a1be-2eccbc491e45", 00:08:47.575 "strip_size_kb": 64, 00:08:47.575 "state": "online", 00:08:47.575 "raid_level": "concat", 00:08:47.575 "superblock": true, 00:08:47.575 "num_base_bdevs": 3, 00:08:47.575 "num_base_bdevs_discovered": 3, 00:08:47.575 "num_base_bdevs_operational": 3, 00:08:47.575 "base_bdevs_list": [ 00:08:47.575 { 00:08:47.575 "name": "BaseBdev1", 00:08:47.575 "uuid": "e6e74bca-409a-5cbe-b0b6-fe983d00b4cd", 00:08:47.575 "is_configured": true, 00:08:47.575 "data_offset": 2048, 00:08:47.575 "data_size": 63488 00:08:47.575 }, 00:08:47.575 { 00:08:47.575 "name": "BaseBdev2", 00:08:47.575 "uuid": "fd760db0-9c56-5eda-bda2-2c0dd53c42f9", 00:08:47.575 "is_configured": true, 00:08:47.575 "data_offset": 2048, 00:08:47.575 "data_size": 63488 00:08:47.575 }, 00:08:47.575 { 00:08:47.575 "name": "BaseBdev3", 00:08:47.575 "uuid": "c2b3faec-19b8-55c7-a69b-0b0160aa7064", 00:08:47.575 "is_configured": true, 00:08:47.575 "data_offset": 2048, 00:08:47.575 "data_size": 63488 00:08:47.575 } 00:08:47.575 ] 00:08:47.575 }' 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:47.575 16:47:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.834 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:47.834 16:47:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:48.093 [2024-09-29 16:47:09.579567] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.031 "name": "raid_bdev1", 00:08:49.031 "uuid": "90a7ca6c-1b1a-4047-a1be-2eccbc491e45", 00:08:49.031 "strip_size_kb": 64, 00:08:49.031 "state": "online", 00:08:49.031 "raid_level": "concat", 00:08:49.031 "superblock": true, 00:08:49.031 "num_base_bdevs": 3, 00:08:49.031 "num_base_bdevs_discovered": 3, 00:08:49.031 "num_base_bdevs_operational": 3, 00:08:49.031 "base_bdevs_list": [ 00:08:49.031 { 00:08:49.031 "name": "BaseBdev1", 00:08:49.031 "uuid": "e6e74bca-409a-5cbe-b0b6-fe983d00b4cd", 00:08:49.031 "is_configured": true, 00:08:49.031 "data_offset": 2048, 00:08:49.031 "data_size": 63488 00:08:49.031 }, 00:08:49.031 { 00:08:49.031 "name": "BaseBdev2", 00:08:49.031 "uuid": "fd760db0-9c56-5eda-bda2-2c0dd53c42f9", 00:08:49.031 "is_configured": true, 00:08:49.031 "data_offset": 2048, 00:08:49.031 "data_size": 63488 00:08:49.031 }, 00:08:49.031 { 00:08:49.031 "name": "BaseBdev3", 00:08:49.031 "uuid": "c2b3faec-19b8-55c7-a69b-0b0160aa7064", 00:08:49.031 "is_configured": true, 00:08:49.031 "data_offset": 2048, 00:08:49.031 "data_size": 63488 00:08:49.031 } 00:08:49.031 ] 00:08:49.031 }' 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.031 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.290 [2024-09-29 16:47:10.926939] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:49.290 [2024-09-29 16:47:10.926978] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:49.290 [2024-09-29 16:47:10.929490] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:49.290 [2024-09-29 16:47:10.929545] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:49.290 [2024-09-29 16:47:10.929577] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:49.290 [2024-09-29 16:47:10.929600] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:49.290 { 00:08:49.290 "results": [ 00:08:49.290 { 00:08:49.290 "job": "raid_bdev1", 00:08:49.290 "core_mask": "0x1", 00:08:49.290 "workload": "randrw", 00:08:49.290 "percentage": 50, 00:08:49.290 "status": "finished", 00:08:49.290 "queue_depth": 1, 00:08:49.290 "io_size": 131072, 00:08:49.290 "runtime": 1.348185, 00:08:49.290 "iops": 17372.986645007917, 00:08:49.290 "mibps": 2171.6233306259896, 00:08:49.290 "io_failed": 1, 00:08:49.290 "io_timeout": 0, 00:08:49.290 "avg_latency_us": 79.69875882455699, 00:08:49.290 "min_latency_us": 24.258515283842794, 00:08:49.290 "max_latency_us": 1366.5257641921398 00:08:49.290 } 00:08:49.290 ], 00:08:49.290 "core_count": 1 00:08:49.290 } 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78031 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 78031 ']' 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 78031 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78031 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:49.290 killing process with pid 78031 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78031' 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 78031 00:08:49.290 [2024-09-29 16:47:10.958386] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:49.290 16:47:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 78031 00:08:49.550 [2024-09-29 16:47:10.983992] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:49.550 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:49.550 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.C19XyUkWLv 00:08:49.550 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:49.809 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:08:49.809 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:49.809 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:49.809 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:49.809 16:47:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:08:49.809 00:08:49.809 real 0m3.201s 00:08:49.809 user 0m3.990s 00:08:49.809 sys 0m0.511s 00:08:49.809 16:47:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:49.809 16:47:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.809 ************************************ 00:08:49.809 END TEST raid_write_error_test 00:08:49.809 ************************************ 00:08:49.809 16:47:11 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:49.809 16:47:11 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:08:49.809 16:47:11 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:49.809 16:47:11 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:49.809 16:47:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:49.809 ************************************ 00:08:49.809 START TEST raid_state_function_test 00:08:49.809 ************************************ 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 false 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=78158 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:49.809 Process raid pid: 78158 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78158' 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 78158 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 78158 ']' 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:49.809 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:49.809 16:47:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.809 [2024-09-29 16:47:11.386478] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:49.810 [2024-09-29 16:47:11.386616] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:50.069 [2024-09-29 16:47:11.531856] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:50.069 [2024-09-29 16:47:11.577381] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:50.069 [2024-09-29 16:47:11.619079] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:50.069 [2024-09-29 16:47:11.619116] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.637 [2024-09-29 16:47:12.208179] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:50.637 [2024-09-29 16:47:12.208234] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:50.637 [2024-09-29 16:47:12.208247] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:50.637 [2024-09-29 16:47:12.208258] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:50.637 [2024-09-29 16:47:12.208264] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:50.637 [2024-09-29 16:47:12.208274] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.637 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.638 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.638 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.638 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.638 "name": "Existed_Raid", 00:08:50.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.638 "strip_size_kb": 0, 00:08:50.638 "state": "configuring", 00:08:50.638 "raid_level": "raid1", 00:08:50.638 "superblock": false, 00:08:50.638 "num_base_bdevs": 3, 00:08:50.638 "num_base_bdevs_discovered": 0, 00:08:50.638 "num_base_bdevs_operational": 3, 00:08:50.638 "base_bdevs_list": [ 00:08:50.638 { 00:08:50.638 "name": "BaseBdev1", 00:08:50.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.638 "is_configured": false, 00:08:50.638 "data_offset": 0, 00:08:50.638 "data_size": 0 00:08:50.638 }, 00:08:50.638 { 00:08:50.638 "name": "BaseBdev2", 00:08:50.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.638 "is_configured": false, 00:08:50.638 "data_offset": 0, 00:08:50.638 "data_size": 0 00:08:50.638 }, 00:08:50.638 { 00:08:50.638 "name": "BaseBdev3", 00:08:50.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:50.638 "is_configured": false, 00:08:50.638 "data_offset": 0, 00:08:50.638 "data_size": 0 00:08:50.638 } 00:08:50.638 ] 00:08:50.638 }' 00:08:50.638 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.638 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.205 [2024-09-29 16:47:12.635348] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:51.205 [2024-09-29 16:47:12.635398] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.205 [2024-09-29 16:47:12.647332] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:51.205 [2024-09-29 16:47:12.647374] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:51.205 [2024-09-29 16:47:12.647382] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:51.205 [2024-09-29 16:47:12.647391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:51.205 [2024-09-29 16:47:12.647397] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:51.205 [2024-09-29 16:47:12.647413] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.205 [2024-09-29 16:47:12.668153] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:51.205 BaseBdev1 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.205 [ 00:08:51.205 { 00:08:51.205 "name": "BaseBdev1", 00:08:51.205 "aliases": [ 00:08:51.205 "6ba187f9-49ad-495c-958e-52662e68ff9c" 00:08:51.205 ], 00:08:51.205 "product_name": "Malloc disk", 00:08:51.205 "block_size": 512, 00:08:51.205 "num_blocks": 65536, 00:08:51.205 "uuid": "6ba187f9-49ad-495c-958e-52662e68ff9c", 00:08:51.205 "assigned_rate_limits": { 00:08:51.205 "rw_ios_per_sec": 0, 00:08:51.205 "rw_mbytes_per_sec": 0, 00:08:51.205 "r_mbytes_per_sec": 0, 00:08:51.205 "w_mbytes_per_sec": 0 00:08:51.205 }, 00:08:51.205 "claimed": true, 00:08:51.205 "claim_type": "exclusive_write", 00:08:51.205 "zoned": false, 00:08:51.205 "supported_io_types": { 00:08:51.205 "read": true, 00:08:51.205 "write": true, 00:08:51.205 "unmap": true, 00:08:51.205 "flush": true, 00:08:51.205 "reset": true, 00:08:51.205 "nvme_admin": false, 00:08:51.205 "nvme_io": false, 00:08:51.205 "nvme_io_md": false, 00:08:51.205 "write_zeroes": true, 00:08:51.205 "zcopy": true, 00:08:51.205 "get_zone_info": false, 00:08:51.205 "zone_management": false, 00:08:51.205 "zone_append": false, 00:08:51.205 "compare": false, 00:08:51.205 "compare_and_write": false, 00:08:51.205 "abort": true, 00:08:51.205 "seek_hole": false, 00:08:51.205 "seek_data": false, 00:08:51.205 "copy": true, 00:08:51.205 "nvme_iov_md": false 00:08:51.205 }, 00:08:51.205 "memory_domains": [ 00:08:51.205 { 00:08:51.205 "dma_device_id": "system", 00:08:51.205 "dma_device_type": 1 00:08:51.205 }, 00:08:51.205 { 00:08:51.205 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.205 "dma_device_type": 2 00:08:51.205 } 00:08:51.205 ], 00:08:51.205 "driver_specific": {} 00:08:51.205 } 00:08:51.205 ] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.205 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.205 "name": "Existed_Raid", 00:08:51.205 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.205 "strip_size_kb": 0, 00:08:51.206 "state": "configuring", 00:08:51.206 "raid_level": "raid1", 00:08:51.206 "superblock": false, 00:08:51.206 "num_base_bdevs": 3, 00:08:51.206 "num_base_bdevs_discovered": 1, 00:08:51.206 "num_base_bdevs_operational": 3, 00:08:51.206 "base_bdevs_list": [ 00:08:51.206 { 00:08:51.206 "name": "BaseBdev1", 00:08:51.206 "uuid": "6ba187f9-49ad-495c-958e-52662e68ff9c", 00:08:51.206 "is_configured": true, 00:08:51.206 "data_offset": 0, 00:08:51.206 "data_size": 65536 00:08:51.206 }, 00:08:51.206 { 00:08:51.206 "name": "BaseBdev2", 00:08:51.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.206 "is_configured": false, 00:08:51.206 "data_offset": 0, 00:08:51.206 "data_size": 0 00:08:51.206 }, 00:08:51.206 { 00:08:51.206 "name": "BaseBdev3", 00:08:51.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.206 "is_configured": false, 00:08:51.206 "data_offset": 0, 00:08:51.206 "data_size": 0 00:08:51.206 } 00:08:51.206 ] 00:08:51.206 }' 00:08:51.206 16:47:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.206 16:47:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.464 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:51.464 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.464 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.464 [2024-09-29 16:47:13.131382] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:51.464 [2024-09-29 16:47:13.131430] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:51.464 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.464 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:51.464 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.465 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.723 [2024-09-29 16:47:13.143413] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:51.723 [2024-09-29 16:47:13.145270] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:51.723 [2024-09-29 16:47:13.145326] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:51.723 [2024-09-29 16:47:13.145336] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:51.723 [2024-09-29 16:47:13.145346] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:51.723 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.723 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:51.723 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:51.723 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:51.723 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:51.724 "name": "Existed_Raid", 00:08:51.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.724 "strip_size_kb": 0, 00:08:51.724 "state": "configuring", 00:08:51.724 "raid_level": "raid1", 00:08:51.724 "superblock": false, 00:08:51.724 "num_base_bdevs": 3, 00:08:51.724 "num_base_bdevs_discovered": 1, 00:08:51.724 "num_base_bdevs_operational": 3, 00:08:51.724 "base_bdevs_list": [ 00:08:51.724 { 00:08:51.724 "name": "BaseBdev1", 00:08:51.724 "uuid": "6ba187f9-49ad-495c-958e-52662e68ff9c", 00:08:51.724 "is_configured": true, 00:08:51.724 "data_offset": 0, 00:08:51.724 "data_size": 65536 00:08:51.724 }, 00:08:51.724 { 00:08:51.724 "name": "BaseBdev2", 00:08:51.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.724 "is_configured": false, 00:08:51.724 "data_offset": 0, 00:08:51.724 "data_size": 0 00:08:51.724 }, 00:08:51.724 { 00:08:51.724 "name": "BaseBdev3", 00:08:51.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:51.724 "is_configured": false, 00:08:51.724 "data_offset": 0, 00:08:51.724 "data_size": 0 00:08:51.724 } 00:08:51.724 ] 00:08:51.724 }' 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:51.724 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.984 [2024-09-29 16:47:13.592268] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:51.984 BaseBdev2 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.984 [ 00:08:51.984 { 00:08:51.984 "name": "BaseBdev2", 00:08:51.984 "aliases": [ 00:08:51.984 "74cd0721-ac43-48df-992a-44ecffbb4590" 00:08:51.984 ], 00:08:51.984 "product_name": "Malloc disk", 00:08:51.984 "block_size": 512, 00:08:51.984 "num_blocks": 65536, 00:08:51.984 "uuid": "74cd0721-ac43-48df-992a-44ecffbb4590", 00:08:51.984 "assigned_rate_limits": { 00:08:51.984 "rw_ios_per_sec": 0, 00:08:51.984 "rw_mbytes_per_sec": 0, 00:08:51.984 "r_mbytes_per_sec": 0, 00:08:51.984 "w_mbytes_per_sec": 0 00:08:51.984 }, 00:08:51.984 "claimed": true, 00:08:51.984 "claim_type": "exclusive_write", 00:08:51.984 "zoned": false, 00:08:51.984 "supported_io_types": { 00:08:51.984 "read": true, 00:08:51.984 "write": true, 00:08:51.984 "unmap": true, 00:08:51.984 "flush": true, 00:08:51.984 "reset": true, 00:08:51.984 "nvme_admin": false, 00:08:51.984 "nvme_io": false, 00:08:51.984 "nvme_io_md": false, 00:08:51.984 "write_zeroes": true, 00:08:51.984 "zcopy": true, 00:08:51.984 "get_zone_info": false, 00:08:51.984 "zone_management": false, 00:08:51.984 "zone_append": false, 00:08:51.984 "compare": false, 00:08:51.984 "compare_and_write": false, 00:08:51.984 "abort": true, 00:08:51.984 "seek_hole": false, 00:08:51.984 "seek_data": false, 00:08:51.984 "copy": true, 00:08:51.984 "nvme_iov_md": false 00:08:51.984 }, 00:08:51.984 "memory_domains": [ 00:08:51.984 { 00:08:51.984 "dma_device_id": "system", 00:08:51.984 "dma_device_type": 1 00:08:51.984 }, 00:08:51.984 { 00:08:51.984 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:51.984 "dma_device_type": 2 00:08:51.984 } 00:08:51.984 ], 00:08:51.984 "driver_specific": {} 00:08:51.984 } 00:08:51.984 ] 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:51.984 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.243 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.244 "name": "Existed_Raid", 00:08:52.244 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.244 "strip_size_kb": 0, 00:08:52.244 "state": "configuring", 00:08:52.244 "raid_level": "raid1", 00:08:52.244 "superblock": false, 00:08:52.244 "num_base_bdevs": 3, 00:08:52.244 "num_base_bdevs_discovered": 2, 00:08:52.244 "num_base_bdevs_operational": 3, 00:08:52.244 "base_bdevs_list": [ 00:08:52.244 { 00:08:52.244 "name": "BaseBdev1", 00:08:52.244 "uuid": "6ba187f9-49ad-495c-958e-52662e68ff9c", 00:08:52.244 "is_configured": true, 00:08:52.244 "data_offset": 0, 00:08:52.244 "data_size": 65536 00:08:52.244 }, 00:08:52.244 { 00:08:52.244 "name": "BaseBdev2", 00:08:52.244 "uuid": "74cd0721-ac43-48df-992a-44ecffbb4590", 00:08:52.244 "is_configured": true, 00:08:52.244 "data_offset": 0, 00:08:52.244 "data_size": 65536 00:08:52.244 }, 00:08:52.244 { 00:08:52.244 "name": "BaseBdev3", 00:08:52.244 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:52.244 "is_configured": false, 00:08:52.244 "data_offset": 0, 00:08:52.244 "data_size": 0 00:08:52.244 } 00:08:52.244 ] 00:08:52.244 }' 00:08:52.244 16:47:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.244 16:47:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.503 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:52.503 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.503 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.503 [2024-09-29 16:47:14.078493] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:52.503 [2024-09-29 16:47:14.078542] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:52.504 [2024-09-29 16:47:14.078568] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:52.504 [2024-09-29 16:47:14.078880] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:52.504 [2024-09-29 16:47:14.079129] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:52.504 [2024-09-29 16:47:14.079154] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:52.504 [2024-09-29 16:47:14.079369] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:52.504 BaseBdev3 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.504 [ 00:08:52.504 { 00:08:52.504 "name": "BaseBdev3", 00:08:52.504 "aliases": [ 00:08:52.504 "b1b8750c-a6d6-4148-8420-93c0a333aa5d" 00:08:52.504 ], 00:08:52.504 "product_name": "Malloc disk", 00:08:52.504 "block_size": 512, 00:08:52.504 "num_blocks": 65536, 00:08:52.504 "uuid": "b1b8750c-a6d6-4148-8420-93c0a333aa5d", 00:08:52.504 "assigned_rate_limits": { 00:08:52.504 "rw_ios_per_sec": 0, 00:08:52.504 "rw_mbytes_per_sec": 0, 00:08:52.504 "r_mbytes_per_sec": 0, 00:08:52.504 "w_mbytes_per_sec": 0 00:08:52.504 }, 00:08:52.504 "claimed": true, 00:08:52.504 "claim_type": "exclusive_write", 00:08:52.504 "zoned": false, 00:08:52.504 "supported_io_types": { 00:08:52.504 "read": true, 00:08:52.504 "write": true, 00:08:52.504 "unmap": true, 00:08:52.504 "flush": true, 00:08:52.504 "reset": true, 00:08:52.504 "nvme_admin": false, 00:08:52.504 "nvme_io": false, 00:08:52.504 "nvme_io_md": false, 00:08:52.504 "write_zeroes": true, 00:08:52.504 "zcopy": true, 00:08:52.504 "get_zone_info": false, 00:08:52.504 "zone_management": false, 00:08:52.504 "zone_append": false, 00:08:52.504 "compare": false, 00:08:52.504 "compare_and_write": false, 00:08:52.504 "abort": true, 00:08:52.504 "seek_hole": false, 00:08:52.504 "seek_data": false, 00:08:52.504 "copy": true, 00:08:52.504 "nvme_iov_md": false 00:08:52.504 }, 00:08:52.504 "memory_domains": [ 00:08:52.504 { 00:08:52.504 "dma_device_id": "system", 00:08:52.504 "dma_device_type": 1 00:08:52.504 }, 00:08:52.504 { 00:08:52.504 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:52.504 "dma_device_type": 2 00:08:52.504 } 00:08:52.504 ], 00:08:52.504 "driver_specific": {} 00:08:52.504 } 00:08:52.504 ] 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.504 "name": "Existed_Raid", 00:08:52.504 "uuid": "dc602ff4-4e66-4ad2-9e60-487090f61663", 00:08:52.504 "strip_size_kb": 0, 00:08:52.504 "state": "online", 00:08:52.504 "raid_level": "raid1", 00:08:52.504 "superblock": false, 00:08:52.504 "num_base_bdevs": 3, 00:08:52.504 "num_base_bdevs_discovered": 3, 00:08:52.504 "num_base_bdevs_operational": 3, 00:08:52.504 "base_bdevs_list": [ 00:08:52.504 { 00:08:52.504 "name": "BaseBdev1", 00:08:52.504 "uuid": "6ba187f9-49ad-495c-958e-52662e68ff9c", 00:08:52.504 "is_configured": true, 00:08:52.504 "data_offset": 0, 00:08:52.504 "data_size": 65536 00:08:52.504 }, 00:08:52.504 { 00:08:52.504 "name": "BaseBdev2", 00:08:52.504 "uuid": "74cd0721-ac43-48df-992a-44ecffbb4590", 00:08:52.504 "is_configured": true, 00:08:52.504 "data_offset": 0, 00:08:52.504 "data_size": 65536 00:08:52.504 }, 00:08:52.504 { 00:08:52.504 "name": "BaseBdev3", 00:08:52.504 "uuid": "b1b8750c-a6d6-4148-8420-93c0a333aa5d", 00:08:52.504 "is_configured": true, 00:08:52.504 "data_offset": 0, 00:08:52.504 "data_size": 65536 00:08:52.504 } 00:08:52.504 ] 00:08:52.504 }' 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.504 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.073 [2024-09-29 16:47:14.530050] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.073 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:53.073 "name": "Existed_Raid", 00:08:53.073 "aliases": [ 00:08:53.073 "dc602ff4-4e66-4ad2-9e60-487090f61663" 00:08:53.073 ], 00:08:53.073 "product_name": "Raid Volume", 00:08:53.073 "block_size": 512, 00:08:53.073 "num_blocks": 65536, 00:08:53.073 "uuid": "dc602ff4-4e66-4ad2-9e60-487090f61663", 00:08:53.073 "assigned_rate_limits": { 00:08:53.073 "rw_ios_per_sec": 0, 00:08:53.073 "rw_mbytes_per_sec": 0, 00:08:53.073 "r_mbytes_per_sec": 0, 00:08:53.073 "w_mbytes_per_sec": 0 00:08:53.073 }, 00:08:53.073 "claimed": false, 00:08:53.073 "zoned": false, 00:08:53.073 "supported_io_types": { 00:08:53.073 "read": true, 00:08:53.073 "write": true, 00:08:53.073 "unmap": false, 00:08:53.073 "flush": false, 00:08:53.073 "reset": true, 00:08:53.073 "nvme_admin": false, 00:08:53.073 "nvme_io": false, 00:08:53.073 "nvme_io_md": false, 00:08:53.073 "write_zeroes": true, 00:08:53.073 "zcopy": false, 00:08:53.073 "get_zone_info": false, 00:08:53.073 "zone_management": false, 00:08:53.073 "zone_append": false, 00:08:53.073 "compare": false, 00:08:53.073 "compare_and_write": false, 00:08:53.073 "abort": false, 00:08:53.073 "seek_hole": false, 00:08:53.073 "seek_data": false, 00:08:53.073 "copy": false, 00:08:53.073 "nvme_iov_md": false 00:08:53.073 }, 00:08:53.073 "memory_domains": [ 00:08:53.073 { 00:08:53.073 "dma_device_id": "system", 00:08:53.073 "dma_device_type": 1 00:08:53.073 }, 00:08:53.073 { 00:08:53.073 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.074 "dma_device_type": 2 00:08:53.074 }, 00:08:53.074 { 00:08:53.074 "dma_device_id": "system", 00:08:53.074 "dma_device_type": 1 00:08:53.074 }, 00:08:53.074 { 00:08:53.074 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.074 "dma_device_type": 2 00:08:53.074 }, 00:08:53.074 { 00:08:53.074 "dma_device_id": "system", 00:08:53.074 "dma_device_type": 1 00:08:53.074 }, 00:08:53.074 { 00:08:53.074 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.074 "dma_device_type": 2 00:08:53.074 } 00:08:53.074 ], 00:08:53.074 "driver_specific": { 00:08:53.074 "raid": { 00:08:53.074 "uuid": "dc602ff4-4e66-4ad2-9e60-487090f61663", 00:08:53.074 "strip_size_kb": 0, 00:08:53.074 "state": "online", 00:08:53.074 "raid_level": "raid1", 00:08:53.074 "superblock": false, 00:08:53.074 "num_base_bdevs": 3, 00:08:53.074 "num_base_bdevs_discovered": 3, 00:08:53.074 "num_base_bdevs_operational": 3, 00:08:53.074 "base_bdevs_list": [ 00:08:53.074 { 00:08:53.074 "name": "BaseBdev1", 00:08:53.074 "uuid": "6ba187f9-49ad-495c-958e-52662e68ff9c", 00:08:53.074 "is_configured": true, 00:08:53.074 "data_offset": 0, 00:08:53.074 "data_size": 65536 00:08:53.074 }, 00:08:53.074 { 00:08:53.074 "name": "BaseBdev2", 00:08:53.074 "uuid": "74cd0721-ac43-48df-992a-44ecffbb4590", 00:08:53.074 "is_configured": true, 00:08:53.074 "data_offset": 0, 00:08:53.074 "data_size": 65536 00:08:53.074 }, 00:08:53.074 { 00:08:53.074 "name": "BaseBdev3", 00:08:53.074 "uuid": "b1b8750c-a6d6-4148-8420-93c0a333aa5d", 00:08:53.074 "is_configured": true, 00:08:53.074 "data_offset": 0, 00:08:53.074 "data_size": 65536 00:08:53.074 } 00:08:53.074 ] 00:08:53.074 } 00:08:53.074 } 00:08:53.074 }' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:53.074 BaseBdev2 00:08:53.074 BaseBdev3' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:53.074 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:53.333 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.334 [2024-09-29 16:47:14.797343] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:53.334 "name": "Existed_Raid", 00:08:53.334 "uuid": "dc602ff4-4e66-4ad2-9e60-487090f61663", 00:08:53.334 "strip_size_kb": 0, 00:08:53.334 "state": "online", 00:08:53.334 "raid_level": "raid1", 00:08:53.334 "superblock": false, 00:08:53.334 "num_base_bdevs": 3, 00:08:53.334 "num_base_bdevs_discovered": 2, 00:08:53.334 "num_base_bdevs_operational": 2, 00:08:53.334 "base_bdevs_list": [ 00:08:53.334 { 00:08:53.334 "name": null, 00:08:53.334 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:53.334 "is_configured": false, 00:08:53.334 "data_offset": 0, 00:08:53.334 "data_size": 65536 00:08:53.334 }, 00:08:53.334 { 00:08:53.334 "name": "BaseBdev2", 00:08:53.334 "uuid": "74cd0721-ac43-48df-992a-44ecffbb4590", 00:08:53.334 "is_configured": true, 00:08:53.334 "data_offset": 0, 00:08:53.334 "data_size": 65536 00:08:53.334 }, 00:08:53.334 { 00:08:53.334 "name": "BaseBdev3", 00:08:53.334 "uuid": "b1b8750c-a6d6-4148-8420-93c0a333aa5d", 00:08:53.334 "is_configured": true, 00:08:53.334 "data_offset": 0, 00:08:53.334 "data_size": 65536 00:08:53.334 } 00:08:53.334 ] 00:08:53.334 }' 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:53.334 16:47:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.593 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:53.593 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:53.593 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.593 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:53.593 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.593 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 [2024-09-29 16:47:15.303765] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 [2024-09-29 16:47:15.374892] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:53.854 [2024-09-29 16:47:15.374978] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:53.854 [2024-09-29 16:47:15.386113] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:53.854 [2024-09-29 16:47:15.386157] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:53.854 [2024-09-29 16:47:15.386171] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 BaseBdev2 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.854 [ 00:08:53.854 { 00:08:53.854 "name": "BaseBdev2", 00:08:53.854 "aliases": [ 00:08:53.854 "a7246e1e-c99a-4e88-90e8-20238b93f83a" 00:08:53.854 ], 00:08:53.854 "product_name": "Malloc disk", 00:08:53.854 "block_size": 512, 00:08:53.854 "num_blocks": 65536, 00:08:53.854 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:53.854 "assigned_rate_limits": { 00:08:53.854 "rw_ios_per_sec": 0, 00:08:53.854 "rw_mbytes_per_sec": 0, 00:08:53.854 "r_mbytes_per_sec": 0, 00:08:53.854 "w_mbytes_per_sec": 0 00:08:53.854 }, 00:08:53.854 "claimed": false, 00:08:53.854 "zoned": false, 00:08:53.854 "supported_io_types": { 00:08:53.854 "read": true, 00:08:53.854 "write": true, 00:08:53.854 "unmap": true, 00:08:53.854 "flush": true, 00:08:53.854 "reset": true, 00:08:53.854 "nvme_admin": false, 00:08:53.854 "nvme_io": false, 00:08:53.854 "nvme_io_md": false, 00:08:53.854 "write_zeroes": true, 00:08:53.854 "zcopy": true, 00:08:53.854 "get_zone_info": false, 00:08:53.854 "zone_management": false, 00:08:53.854 "zone_append": false, 00:08:53.854 "compare": false, 00:08:53.854 "compare_and_write": false, 00:08:53.854 "abort": true, 00:08:53.854 "seek_hole": false, 00:08:53.854 "seek_data": false, 00:08:53.854 "copy": true, 00:08:53.854 "nvme_iov_md": false 00:08:53.854 }, 00:08:53.854 "memory_domains": [ 00:08:53.854 { 00:08:53.854 "dma_device_id": "system", 00:08:53.854 "dma_device_type": 1 00:08:53.854 }, 00:08:53.854 { 00:08:53.854 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:53.854 "dma_device_type": 2 00:08:53.854 } 00:08:53.854 ], 00:08:53.854 "driver_specific": {} 00:08:53.854 } 00:08:53.854 ] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:53.854 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.855 BaseBdev3 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:53.855 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.115 [ 00:08:54.115 { 00:08:54.115 "name": "BaseBdev3", 00:08:54.115 "aliases": [ 00:08:54.115 "957096e1-ff6a-4b30-984e-9b7cb6626f94" 00:08:54.115 ], 00:08:54.115 "product_name": "Malloc disk", 00:08:54.115 "block_size": 512, 00:08:54.115 "num_blocks": 65536, 00:08:54.115 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:54.115 "assigned_rate_limits": { 00:08:54.115 "rw_ios_per_sec": 0, 00:08:54.115 "rw_mbytes_per_sec": 0, 00:08:54.115 "r_mbytes_per_sec": 0, 00:08:54.115 "w_mbytes_per_sec": 0 00:08:54.115 }, 00:08:54.115 "claimed": false, 00:08:54.115 "zoned": false, 00:08:54.115 "supported_io_types": { 00:08:54.115 "read": true, 00:08:54.115 "write": true, 00:08:54.115 "unmap": true, 00:08:54.115 "flush": true, 00:08:54.115 "reset": true, 00:08:54.115 "nvme_admin": false, 00:08:54.115 "nvme_io": false, 00:08:54.115 "nvme_io_md": false, 00:08:54.115 "write_zeroes": true, 00:08:54.115 "zcopy": true, 00:08:54.115 "get_zone_info": false, 00:08:54.115 "zone_management": false, 00:08:54.115 "zone_append": false, 00:08:54.115 "compare": false, 00:08:54.115 "compare_and_write": false, 00:08:54.115 "abort": true, 00:08:54.115 "seek_hole": false, 00:08:54.115 "seek_data": false, 00:08:54.115 "copy": true, 00:08:54.115 "nvme_iov_md": false 00:08:54.115 }, 00:08:54.115 "memory_domains": [ 00:08:54.115 { 00:08:54.115 "dma_device_id": "system", 00:08:54.115 "dma_device_type": 1 00:08:54.115 }, 00:08:54.115 { 00:08:54.115 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.115 "dma_device_type": 2 00:08:54.115 } 00:08:54.115 ], 00:08:54.115 "driver_specific": {} 00:08:54.115 } 00:08:54.115 ] 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.115 [2024-09-29 16:47:15.545645] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:54.115 [2024-09-29 16:47:15.545733] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:54.115 [2024-09-29 16:47:15.545770] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:54.115 [2024-09-29 16:47:15.547568] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.115 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.115 "name": "Existed_Raid", 00:08:54.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.115 "strip_size_kb": 0, 00:08:54.115 "state": "configuring", 00:08:54.115 "raid_level": "raid1", 00:08:54.115 "superblock": false, 00:08:54.115 "num_base_bdevs": 3, 00:08:54.115 "num_base_bdevs_discovered": 2, 00:08:54.115 "num_base_bdevs_operational": 3, 00:08:54.115 "base_bdevs_list": [ 00:08:54.115 { 00:08:54.115 "name": "BaseBdev1", 00:08:54.115 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.115 "is_configured": false, 00:08:54.115 "data_offset": 0, 00:08:54.115 "data_size": 0 00:08:54.115 }, 00:08:54.115 { 00:08:54.115 "name": "BaseBdev2", 00:08:54.115 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:54.115 "is_configured": true, 00:08:54.115 "data_offset": 0, 00:08:54.115 "data_size": 65536 00:08:54.115 }, 00:08:54.115 { 00:08:54.115 "name": "BaseBdev3", 00:08:54.115 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:54.115 "is_configured": true, 00:08:54.115 "data_offset": 0, 00:08:54.115 "data_size": 65536 00:08:54.115 } 00:08:54.115 ] 00:08:54.115 }' 00:08:54.116 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.116 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.376 [2024-09-29 16:47:15.968876] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.376 16:47:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.376 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.376 "name": "Existed_Raid", 00:08:54.376 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.376 "strip_size_kb": 0, 00:08:54.376 "state": "configuring", 00:08:54.376 "raid_level": "raid1", 00:08:54.376 "superblock": false, 00:08:54.376 "num_base_bdevs": 3, 00:08:54.376 "num_base_bdevs_discovered": 1, 00:08:54.376 "num_base_bdevs_operational": 3, 00:08:54.376 "base_bdevs_list": [ 00:08:54.376 { 00:08:54.376 "name": "BaseBdev1", 00:08:54.376 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.376 "is_configured": false, 00:08:54.376 "data_offset": 0, 00:08:54.376 "data_size": 0 00:08:54.376 }, 00:08:54.376 { 00:08:54.376 "name": null, 00:08:54.376 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:54.376 "is_configured": false, 00:08:54.376 "data_offset": 0, 00:08:54.376 "data_size": 65536 00:08:54.376 }, 00:08:54.376 { 00:08:54.376 "name": "BaseBdev3", 00:08:54.376 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:54.376 "is_configured": true, 00:08:54.376 "data_offset": 0, 00:08:54.376 "data_size": 65536 00:08:54.376 } 00:08:54.376 ] 00:08:54.376 }' 00:08:54.376 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.376 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.947 BaseBdev1 00:08:54.947 [2024-09-29 16:47:16.462901] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.947 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.947 [ 00:08:54.947 { 00:08:54.947 "name": "BaseBdev1", 00:08:54.947 "aliases": [ 00:08:54.947 "013a4c62-2637-4d23-8b5d-82f8188a874a" 00:08:54.947 ], 00:08:54.947 "product_name": "Malloc disk", 00:08:54.947 "block_size": 512, 00:08:54.947 "num_blocks": 65536, 00:08:54.947 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:54.947 "assigned_rate_limits": { 00:08:54.947 "rw_ios_per_sec": 0, 00:08:54.947 "rw_mbytes_per_sec": 0, 00:08:54.947 "r_mbytes_per_sec": 0, 00:08:54.947 "w_mbytes_per_sec": 0 00:08:54.947 }, 00:08:54.947 "claimed": true, 00:08:54.947 "claim_type": "exclusive_write", 00:08:54.947 "zoned": false, 00:08:54.947 "supported_io_types": { 00:08:54.947 "read": true, 00:08:54.947 "write": true, 00:08:54.947 "unmap": true, 00:08:54.947 "flush": true, 00:08:54.947 "reset": true, 00:08:54.947 "nvme_admin": false, 00:08:54.947 "nvme_io": false, 00:08:54.947 "nvme_io_md": false, 00:08:54.947 "write_zeroes": true, 00:08:54.947 "zcopy": true, 00:08:54.947 "get_zone_info": false, 00:08:54.947 "zone_management": false, 00:08:54.947 "zone_append": false, 00:08:54.947 "compare": false, 00:08:54.947 "compare_and_write": false, 00:08:54.947 "abort": true, 00:08:54.947 "seek_hole": false, 00:08:54.947 "seek_data": false, 00:08:54.947 "copy": true, 00:08:54.947 "nvme_iov_md": false 00:08:54.947 }, 00:08:54.948 "memory_domains": [ 00:08:54.948 { 00:08:54.948 "dma_device_id": "system", 00:08:54.948 "dma_device_type": 1 00:08:54.948 }, 00:08:54.948 { 00:08:54.948 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:54.948 "dma_device_type": 2 00:08:54.948 } 00:08:54.948 ], 00:08:54.948 "driver_specific": {} 00:08:54.948 } 00:08:54.948 ] 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.948 "name": "Existed_Raid", 00:08:54.948 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:54.948 "strip_size_kb": 0, 00:08:54.948 "state": "configuring", 00:08:54.948 "raid_level": "raid1", 00:08:54.948 "superblock": false, 00:08:54.948 "num_base_bdevs": 3, 00:08:54.948 "num_base_bdevs_discovered": 2, 00:08:54.948 "num_base_bdevs_operational": 3, 00:08:54.948 "base_bdevs_list": [ 00:08:54.948 { 00:08:54.948 "name": "BaseBdev1", 00:08:54.948 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:54.948 "is_configured": true, 00:08:54.948 "data_offset": 0, 00:08:54.948 "data_size": 65536 00:08:54.948 }, 00:08:54.948 { 00:08:54.948 "name": null, 00:08:54.948 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:54.948 "is_configured": false, 00:08:54.948 "data_offset": 0, 00:08:54.948 "data_size": 65536 00:08:54.948 }, 00:08:54.948 { 00:08:54.948 "name": "BaseBdev3", 00:08:54.948 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:54.948 "is_configured": true, 00:08:54.948 "data_offset": 0, 00:08:54.948 "data_size": 65536 00:08:54.948 } 00:08:54.948 ] 00:08:54.948 }' 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.948 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.518 [2024-09-29 16:47:16.958124] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:55.518 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.519 16:47:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.519 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:55.519 "name": "Existed_Raid", 00:08:55.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:55.519 "strip_size_kb": 0, 00:08:55.519 "state": "configuring", 00:08:55.519 "raid_level": "raid1", 00:08:55.519 "superblock": false, 00:08:55.519 "num_base_bdevs": 3, 00:08:55.519 "num_base_bdevs_discovered": 1, 00:08:55.519 "num_base_bdevs_operational": 3, 00:08:55.519 "base_bdevs_list": [ 00:08:55.519 { 00:08:55.519 "name": "BaseBdev1", 00:08:55.519 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:55.519 "is_configured": true, 00:08:55.519 "data_offset": 0, 00:08:55.519 "data_size": 65536 00:08:55.519 }, 00:08:55.519 { 00:08:55.519 "name": null, 00:08:55.519 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:55.519 "is_configured": false, 00:08:55.519 "data_offset": 0, 00:08:55.519 "data_size": 65536 00:08:55.519 }, 00:08:55.519 { 00:08:55.519 "name": null, 00:08:55.519 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:55.519 "is_configured": false, 00:08:55.519 "data_offset": 0, 00:08:55.519 "data_size": 65536 00:08:55.519 } 00:08:55.519 ] 00:08:55.519 }' 00:08:55.519 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:55.519 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:55.779 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.779 [2024-09-29 16:47:17.449311] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.039 "name": "Existed_Raid", 00:08:56.039 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.039 "strip_size_kb": 0, 00:08:56.039 "state": "configuring", 00:08:56.039 "raid_level": "raid1", 00:08:56.039 "superblock": false, 00:08:56.039 "num_base_bdevs": 3, 00:08:56.039 "num_base_bdevs_discovered": 2, 00:08:56.039 "num_base_bdevs_operational": 3, 00:08:56.039 "base_bdevs_list": [ 00:08:56.039 { 00:08:56.039 "name": "BaseBdev1", 00:08:56.039 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:56.039 "is_configured": true, 00:08:56.039 "data_offset": 0, 00:08:56.039 "data_size": 65536 00:08:56.039 }, 00:08:56.039 { 00:08:56.039 "name": null, 00:08:56.039 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:56.039 "is_configured": false, 00:08:56.039 "data_offset": 0, 00:08:56.039 "data_size": 65536 00:08:56.039 }, 00:08:56.039 { 00:08:56.039 "name": "BaseBdev3", 00:08:56.039 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:56.039 "is_configured": true, 00:08:56.039 "data_offset": 0, 00:08:56.039 "data_size": 65536 00:08:56.039 } 00:08:56.039 ] 00:08:56.039 }' 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.039 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.371 [2024-09-29 16:47:17.948480] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.371 16:47:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.655 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.655 "name": "Existed_Raid", 00:08:56.655 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.655 "strip_size_kb": 0, 00:08:56.655 "state": "configuring", 00:08:56.655 "raid_level": "raid1", 00:08:56.655 "superblock": false, 00:08:56.655 "num_base_bdevs": 3, 00:08:56.655 "num_base_bdevs_discovered": 1, 00:08:56.655 "num_base_bdevs_operational": 3, 00:08:56.655 "base_bdevs_list": [ 00:08:56.655 { 00:08:56.655 "name": null, 00:08:56.655 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:56.655 "is_configured": false, 00:08:56.655 "data_offset": 0, 00:08:56.655 "data_size": 65536 00:08:56.655 }, 00:08:56.655 { 00:08:56.656 "name": null, 00:08:56.656 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:56.656 "is_configured": false, 00:08:56.656 "data_offset": 0, 00:08:56.656 "data_size": 65536 00:08:56.656 }, 00:08:56.656 { 00:08:56.656 "name": "BaseBdev3", 00:08:56.656 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:56.656 "is_configured": true, 00:08:56.656 "data_offset": 0, 00:08:56.656 "data_size": 65536 00:08:56.656 } 00:08:56.656 ] 00:08:56.656 }' 00:08:56.656 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.656 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.915 [2024-09-29 16:47:18.450157] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.915 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.916 "name": "Existed_Raid", 00:08:56.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.916 "strip_size_kb": 0, 00:08:56.916 "state": "configuring", 00:08:56.916 "raid_level": "raid1", 00:08:56.916 "superblock": false, 00:08:56.916 "num_base_bdevs": 3, 00:08:56.916 "num_base_bdevs_discovered": 2, 00:08:56.916 "num_base_bdevs_operational": 3, 00:08:56.916 "base_bdevs_list": [ 00:08:56.916 { 00:08:56.916 "name": null, 00:08:56.916 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:56.916 "is_configured": false, 00:08:56.916 "data_offset": 0, 00:08:56.916 "data_size": 65536 00:08:56.916 }, 00:08:56.916 { 00:08:56.916 "name": "BaseBdev2", 00:08:56.916 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:56.916 "is_configured": true, 00:08:56.916 "data_offset": 0, 00:08:56.916 "data_size": 65536 00:08:56.916 }, 00:08:56.916 { 00:08:56.916 "name": "BaseBdev3", 00:08:56.916 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:56.916 "is_configured": true, 00:08:56.916 "data_offset": 0, 00:08:56.916 "data_size": 65536 00:08:56.916 } 00:08:56.916 ] 00:08:56.916 }' 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.916 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 013a4c62-2637-4d23-8b5d-82f8188a874a 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.496 [2024-09-29 16:47:18.964195] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:57.496 [2024-09-29 16:47:18.964304] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:57.496 [2024-09-29 16:47:18.964329] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:57.496 [2024-09-29 16:47:18.964630] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:57.496 [2024-09-29 16:47:18.964813] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:57.496 [2024-09-29 16:47:18.964865] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:57.496 [2024-09-29 16:47:18.965091] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:57.496 NewBaseBdev 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.496 16:47:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.496 [ 00:08:57.496 { 00:08:57.496 "name": "NewBaseBdev", 00:08:57.496 "aliases": [ 00:08:57.496 "013a4c62-2637-4d23-8b5d-82f8188a874a" 00:08:57.496 ], 00:08:57.496 "product_name": "Malloc disk", 00:08:57.496 "block_size": 512, 00:08:57.496 "num_blocks": 65536, 00:08:57.496 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:57.496 "assigned_rate_limits": { 00:08:57.496 "rw_ios_per_sec": 0, 00:08:57.496 "rw_mbytes_per_sec": 0, 00:08:57.496 "r_mbytes_per_sec": 0, 00:08:57.496 "w_mbytes_per_sec": 0 00:08:57.496 }, 00:08:57.496 "claimed": true, 00:08:57.496 "claim_type": "exclusive_write", 00:08:57.496 "zoned": false, 00:08:57.496 "supported_io_types": { 00:08:57.496 "read": true, 00:08:57.496 "write": true, 00:08:57.496 "unmap": true, 00:08:57.496 "flush": true, 00:08:57.496 "reset": true, 00:08:57.496 "nvme_admin": false, 00:08:57.496 "nvme_io": false, 00:08:57.496 "nvme_io_md": false, 00:08:57.496 "write_zeroes": true, 00:08:57.496 "zcopy": true, 00:08:57.496 "get_zone_info": false, 00:08:57.496 "zone_management": false, 00:08:57.496 "zone_append": false, 00:08:57.496 "compare": false, 00:08:57.496 "compare_and_write": false, 00:08:57.496 "abort": true, 00:08:57.496 "seek_hole": false, 00:08:57.496 "seek_data": false, 00:08:57.496 "copy": true, 00:08:57.496 "nvme_iov_md": false 00:08:57.496 }, 00:08:57.496 "memory_domains": [ 00:08:57.496 { 00:08:57.497 "dma_device_id": "system", 00:08:57.497 "dma_device_type": 1 00:08:57.497 }, 00:08:57.497 { 00:08:57.497 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.497 "dma_device_type": 2 00:08:57.497 } 00:08:57.497 ], 00:08:57.497 "driver_specific": {} 00:08:57.497 } 00:08:57.497 ] 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.497 "name": "Existed_Raid", 00:08:57.497 "uuid": "356b44a2-8e18-4aa9-8f44-6bcf35c77880", 00:08:57.497 "strip_size_kb": 0, 00:08:57.497 "state": "online", 00:08:57.497 "raid_level": "raid1", 00:08:57.497 "superblock": false, 00:08:57.497 "num_base_bdevs": 3, 00:08:57.497 "num_base_bdevs_discovered": 3, 00:08:57.497 "num_base_bdevs_operational": 3, 00:08:57.497 "base_bdevs_list": [ 00:08:57.497 { 00:08:57.497 "name": "NewBaseBdev", 00:08:57.497 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:57.497 "is_configured": true, 00:08:57.497 "data_offset": 0, 00:08:57.497 "data_size": 65536 00:08:57.497 }, 00:08:57.497 { 00:08:57.497 "name": "BaseBdev2", 00:08:57.497 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:57.497 "is_configured": true, 00:08:57.497 "data_offset": 0, 00:08:57.497 "data_size": 65536 00:08:57.497 }, 00:08:57.497 { 00:08:57.497 "name": "BaseBdev3", 00:08:57.497 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:57.497 "is_configured": true, 00:08:57.497 "data_offset": 0, 00:08:57.497 "data_size": 65536 00:08:57.497 } 00:08:57.497 ] 00:08:57.497 }' 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.497 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.758 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.018 [2024-09-29 16:47:19.431715] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:58.018 "name": "Existed_Raid", 00:08:58.018 "aliases": [ 00:08:58.018 "356b44a2-8e18-4aa9-8f44-6bcf35c77880" 00:08:58.018 ], 00:08:58.018 "product_name": "Raid Volume", 00:08:58.018 "block_size": 512, 00:08:58.018 "num_blocks": 65536, 00:08:58.018 "uuid": "356b44a2-8e18-4aa9-8f44-6bcf35c77880", 00:08:58.018 "assigned_rate_limits": { 00:08:58.018 "rw_ios_per_sec": 0, 00:08:58.018 "rw_mbytes_per_sec": 0, 00:08:58.018 "r_mbytes_per_sec": 0, 00:08:58.018 "w_mbytes_per_sec": 0 00:08:58.018 }, 00:08:58.018 "claimed": false, 00:08:58.018 "zoned": false, 00:08:58.018 "supported_io_types": { 00:08:58.018 "read": true, 00:08:58.018 "write": true, 00:08:58.018 "unmap": false, 00:08:58.018 "flush": false, 00:08:58.018 "reset": true, 00:08:58.018 "nvme_admin": false, 00:08:58.018 "nvme_io": false, 00:08:58.018 "nvme_io_md": false, 00:08:58.018 "write_zeroes": true, 00:08:58.018 "zcopy": false, 00:08:58.018 "get_zone_info": false, 00:08:58.018 "zone_management": false, 00:08:58.018 "zone_append": false, 00:08:58.018 "compare": false, 00:08:58.018 "compare_and_write": false, 00:08:58.018 "abort": false, 00:08:58.018 "seek_hole": false, 00:08:58.018 "seek_data": false, 00:08:58.018 "copy": false, 00:08:58.018 "nvme_iov_md": false 00:08:58.018 }, 00:08:58.018 "memory_domains": [ 00:08:58.018 { 00:08:58.018 "dma_device_id": "system", 00:08:58.018 "dma_device_type": 1 00:08:58.018 }, 00:08:58.018 { 00:08:58.018 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.018 "dma_device_type": 2 00:08:58.018 }, 00:08:58.018 { 00:08:58.018 "dma_device_id": "system", 00:08:58.018 "dma_device_type": 1 00:08:58.018 }, 00:08:58.018 { 00:08:58.018 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.018 "dma_device_type": 2 00:08:58.018 }, 00:08:58.018 { 00:08:58.018 "dma_device_id": "system", 00:08:58.018 "dma_device_type": 1 00:08:58.018 }, 00:08:58.018 { 00:08:58.018 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.018 "dma_device_type": 2 00:08:58.018 } 00:08:58.018 ], 00:08:58.018 "driver_specific": { 00:08:58.018 "raid": { 00:08:58.018 "uuid": "356b44a2-8e18-4aa9-8f44-6bcf35c77880", 00:08:58.018 "strip_size_kb": 0, 00:08:58.018 "state": "online", 00:08:58.018 "raid_level": "raid1", 00:08:58.018 "superblock": false, 00:08:58.018 "num_base_bdevs": 3, 00:08:58.018 "num_base_bdevs_discovered": 3, 00:08:58.018 "num_base_bdevs_operational": 3, 00:08:58.018 "base_bdevs_list": [ 00:08:58.018 { 00:08:58.018 "name": "NewBaseBdev", 00:08:58.018 "uuid": "013a4c62-2637-4d23-8b5d-82f8188a874a", 00:08:58.018 "is_configured": true, 00:08:58.018 "data_offset": 0, 00:08:58.018 "data_size": 65536 00:08:58.018 }, 00:08:58.018 { 00:08:58.018 "name": "BaseBdev2", 00:08:58.018 "uuid": "a7246e1e-c99a-4e88-90e8-20238b93f83a", 00:08:58.018 "is_configured": true, 00:08:58.018 "data_offset": 0, 00:08:58.018 "data_size": 65536 00:08:58.018 }, 00:08:58.018 { 00:08:58.018 "name": "BaseBdev3", 00:08:58.018 "uuid": "957096e1-ff6a-4b30-984e-9b7cb6626f94", 00:08:58.018 "is_configured": true, 00:08:58.018 "data_offset": 0, 00:08:58.018 "data_size": 65536 00:08:58.018 } 00:08:58.018 ] 00:08:58.018 } 00:08:58.018 } 00:08:58.018 }' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:58.018 BaseBdev2 00:08:58.018 BaseBdev3' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.018 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.018 [2024-09-29 16:47:19.663010] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:58.018 [2024-09-29 16:47:19.663095] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:58.018 [2024-09-29 16:47:19.663174] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:58.019 [2024-09-29 16:47:19.663448] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:58.019 [2024-09-29 16:47:19.663499] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:58.019 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.019 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 78158 00:08:58.019 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 78158 ']' 00:08:58.019 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 78158 00:08:58.019 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:58.019 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:58.019 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78158 00:08:58.279 killing process with pid 78158 00:08:58.279 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:58.279 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:58.279 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78158' 00:08:58.279 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 78158 00:08:58.279 [2024-09-29 16:47:19.710008] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:58.279 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 78158 00:08:58.279 [2024-09-29 16:47:19.741041] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:58.539 16:47:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:58.539 00:08:58.539 real 0m8.688s 00:08:58.539 user 0m14.825s 00:08:58.539 sys 0m1.774s 00:08:58.539 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:58.539 ************************************ 00:08:58.539 END TEST raid_state_function_test 00:08:58.539 ************************************ 00:08:58.539 16:47:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.539 16:47:20 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:08:58.539 16:47:20 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:58.539 16:47:20 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:58.539 16:47:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:58.539 ************************************ 00:08:58.539 START TEST raid_state_function_test_sb 00:08:58.539 ************************************ 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 true 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78757 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78757' 00:08:58.539 Process raid pid: 78757 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78757 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 78757 ']' 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:58.539 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:58.539 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:58.539 [2024-09-29 16:47:20.150203] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:08:58.539 [2024-09-29 16:47:20.150334] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:58.799 [2024-09-29 16:47:20.297164] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:58.799 [2024-09-29 16:47:20.343021] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:58.799 [2024-09-29 16:47:20.384614] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:58.799 [2024-09-29 16:47:20.384658] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.368 [2024-09-29 16:47:20.977885] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:59.368 [2024-09-29 16:47:20.977934] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:59.368 [2024-09-29 16:47:20.977947] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:59.368 [2024-09-29 16:47:20.977956] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:59.368 [2024-09-29 16:47:20.977962] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:59.368 [2024-09-29 16:47:20.977973] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.368 16:47:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.368 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.368 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.368 "name": "Existed_Raid", 00:08:59.368 "uuid": "823f2448-cf14-4432-8e52-244807708e2f", 00:08:59.368 "strip_size_kb": 0, 00:08:59.368 "state": "configuring", 00:08:59.368 "raid_level": "raid1", 00:08:59.368 "superblock": true, 00:08:59.368 "num_base_bdevs": 3, 00:08:59.368 "num_base_bdevs_discovered": 0, 00:08:59.368 "num_base_bdevs_operational": 3, 00:08:59.368 "base_bdevs_list": [ 00:08:59.368 { 00:08:59.368 "name": "BaseBdev1", 00:08:59.368 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.368 "is_configured": false, 00:08:59.368 "data_offset": 0, 00:08:59.368 "data_size": 0 00:08:59.368 }, 00:08:59.368 { 00:08:59.368 "name": "BaseBdev2", 00:08:59.368 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.368 "is_configured": false, 00:08:59.368 "data_offset": 0, 00:08:59.368 "data_size": 0 00:08:59.368 }, 00:08:59.368 { 00:08:59.368 "name": "BaseBdev3", 00:08:59.368 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.368 "is_configured": false, 00:08:59.368 "data_offset": 0, 00:08:59.368 "data_size": 0 00:08:59.368 } 00:08:59.368 ] 00:08:59.368 }' 00:08:59.368 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.368 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.937 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.938 [2024-09-29 16:47:21.416994] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:59.938 [2024-09-29 16:47:21.417037] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.938 [2024-09-29 16:47:21.424999] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:59.938 [2024-09-29 16:47:21.425042] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:59.938 [2024-09-29 16:47:21.425049] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:59.938 [2024-09-29 16:47:21.425059] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:59.938 [2024-09-29 16:47:21.425065] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:59.938 [2024-09-29 16:47:21.425073] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.938 [2024-09-29 16:47:21.441915] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:59.938 BaseBdev1 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.938 [ 00:08:59.938 { 00:08:59.938 "name": "BaseBdev1", 00:08:59.938 "aliases": [ 00:08:59.938 "178e76db-ebaa-404f-af9a-47576cd9ab83" 00:08:59.938 ], 00:08:59.938 "product_name": "Malloc disk", 00:08:59.938 "block_size": 512, 00:08:59.938 "num_blocks": 65536, 00:08:59.938 "uuid": "178e76db-ebaa-404f-af9a-47576cd9ab83", 00:08:59.938 "assigned_rate_limits": { 00:08:59.938 "rw_ios_per_sec": 0, 00:08:59.938 "rw_mbytes_per_sec": 0, 00:08:59.938 "r_mbytes_per_sec": 0, 00:08:59.938 "w_mbytes_per_sec": 0 00:08:59.938 }, 00:08:59.938 "claimed": true, 00:08:59.938 "claim_type": "exclusive_write", 00:08:59.938 "zoned": false, 00:08:59.938 "supported_io_types": { 00:08:59.938 "read": true, 00:08:59.938 "write": true, 00:08:59.938 "unmap": true, 00:08:59.938 "flush": true, 00:08:59.938 "reset": true, 00:08:59.938 "nvme_admin": false, 00:08:59.938 "nvme_io": false, 00:08:59.938 "nvme_io_md": false, 00:08:59.938 "write_zeroes": true, 00:08:59.938 "zcopy": true, 00:08:59.938 "get_zone_info": false, 00:08:59.938 "zone_management": false, 00:08:59.938 "zone_append": false, 00:08:59.938 "compare": false, 00:08:59.938 "compare_and_write": false, 00:08:59.938 "abort": true, 00:08:59.938 "seek_hole": false, 00:08:59.938 "seek_data": false, 00:08:59.938 "copy": true, 00:08:59.938 "nvme_iov_md": false 00:08:59.938 }, 00:08:59.938 "memory_domains": [ 00:08:59.938 { 00:08:59.938 "dma_device_id": "system", 00:08:59.938 "dma_device_type": 1 00:08:59.938 }, 00:08:59.938 { 00:08:59.938 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.938 "dma_device_type": 2 00:08:59.938 } 00:08:59.938 ], 00:08:59.938 "driver_specific": {} 00:08:59.938 } 00:08:59.938 ] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.938 "name": "Existed_Raid", 00:08:59.938 "uuid": "c3ccc703-2ad8-4c51-a323-2c1017b7b667", 00:08:59.938 "strip_size_kb": 0, 00:08:59.938 "state": "configuring", 00:08:59.938 "raid_level": "raid1", 00:08:59.938 "superblock": true, 00:08:59.938 "num_base_bdevs": 3, 00:08:59.938 "num_base_bdevs_discovered": 1, 00:08:59.938 "num_base_bdevs_operational": 3, 00:08:59.938 "base_bdevs_list": [ 00:08:59.938 { 00:08:59.938 "name": "BaseBdev1", 00:08:59.938 "uuid": "178e76db-ebaa-404f-af9a-47576cd9ab83", 00:08:59.938 "is_configured": true, 00:08:59.938 "data_offset": 2048, 00:08:59.938 "data_size": 63488 00:08:59.938 }, 00:08:59.938 { 00:08:59.938 "name": "BaseBdev2", 00:08:59.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.938 "is_configured": false, 00:08:59.938 "data_offset": 0, 00:08:59.938 "data_size": 0 00:08:59.938 }, 00:08:59.938 { 00:08:59.938 "name": "BaseBdev3", 00:08:59.938 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.938 "is_configured": false, 00:08:59.938 "data_offset": 0, 00:08:59.938 "data_size": 0 00:08:59.938 } 00:08:59.938 ] 00:08:59.938 }' 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.938 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.507 [2024-09-29 16:47:21.889171] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:00.507 [2024-09-29 16:47:21.889215] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.507 [2024-09-29 16:47:21.901213] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:00.507 [2024-09-29 16:47:21.903018] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:00.507 [2024-09-29 16:47:21.903060] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:00.507 [2024-09-29 16:47:21.903069] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:00.507 [2024-09-29 16:47:21.903080] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.507 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.507 "name": "Existed_Raid", 00:09:00.507 "uuid": "492c9f7b-d766-413e-a0df-af9a8c7a6be4", 00:09:00.507 "strip_size_kb": 0, 00:09:00.507 "state": "configuring", 00:09:00.507 "raid_level": "raid1", 00:09:00.508 "superblock": true, 00:09:00.508 "num_base_bdevs": 3, 00:09:00.508 "num_base_bdevs_discovered": 1, 00:09:00.508 "num_base_bdevs_operational": 3, 00:09:00.508 "base_bdevs_list": [ 00:09:00.508 { 00:09:00.508 "name": "BaseBdev1", 00:09:00.508 "uuid": "178e76db-ebaa-404f-af9a-47576cd9ab83", 00:09:00.508 "is_configured": true, 00:09:00.508 "data_offset": 2048, 00:09:00.508 "data_size": 63488 00:09:00.508 }, 00:09:00.508 { 00:09:00.508 "name": "BaseBdev2", 00:09:00.508 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.508 "is_configured": false, 00:09:00.508 "data_offset": 0, 00:09:00.508 "data_size": 0 00:09:00.508 }, 00:09:00.508 { 00:09:00.508 "name": "BaseBdev3", 00:09:00.508 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.508 "is_configured": false, 00:09:00.508 "data_offset": 0, 00:09:00.508 "data_size": 0 00:09:00.508 } 00:09:00.508 ] 00:09:00.508 }' 00:09:00.508 16:47:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.508 16:47:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.768 [2024-09-29 16:47:22.351148] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:00.768 BaseBdev2 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.768 [ 00:09:00.768 { 00:09:00.768 "name": "BaseBdev2", 00:09:00.768 "aliases": [ 00:09:00.768 "35cfd449-ba8b-40d8-9a8b-90eddabb0560" 00:09:00.768 ], 00:09:00.768 "product_name": "Malloc disk", 00:09:00.768 "block_size": 512, 00:09:00.768 "num_blocks": 65536, 00:09:00.768 "uuid": "35cfd449-ba8b-40d8-9a8b-90eddabb0560", 00:09:00.768 "assigned_rate_limits": { 00:09:00.768 "rw_ios_per_sec": 0, 00:09:00.768 "rw_mbytes_per_sec": 0, 00:09:00.768 "r_mbytes_per_sec": 0, 00:09:00.768 "w_mbytes_per_sec": 0 00:09:00.768 }, 00:09:00.768 "claimed": true, 00:09:00.768 "claim_type": "exclusive_write", 00:09:00.768 "zoned": false, 00:09:00.768 "supported_io_types": { 00:09:00.768 "read": true, 00:09:00.768 "write": true, 00:09:00.768 "unmap": true, 00:09:00.768 "flush": true, 00:09:00.768 "reset": true, 00:09:00.768 "nvme_admin": false, 00:09:00.768 "nvme_io": false, 00:09:00.768 "nvme_io_md": false, 00:09:00.768 "write_zeroes": true, 00:09:00.768 "zcopy": true, 00:09:00.768 "get_zone_info": false, 00:09:00.768 "zone_management": false, 00:09:00.768 "zone_append": false, 00:09:00.768 "compare": false, 00:09:00.768 "compare_and_write": false, 00:09:00.768 "abort": true, 00:09:00.768 "seek_hole": false, 00:09:00.768 "seek_data": false, 00:09:00.768 "copy": true, 00:09:00.768 "nvme_iov_md": false 00:09:00.768 }, 00:09:00.768 "memory_domains": [ 00:09:00.768 { 00:09:00.768 "dma_device_id": "system", 00:09:00.768 "dma_device_type": 1 00:09:00.768 }, 00:09:00.768 { 00:09:00.768 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.768 "dma_device_type": 2 00:09:00.768 } 00:09:00.768 ], 00:09:00.768 "driver_specific": {} 00:09:00.768 } 00:09:00.768 ] 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:00.768 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.028 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.028 "name": "Existed_Raid", 00:09:01.028 "uuid": "492c9f7b-d766-413e-a0df-af9a8c7a6be4", 00:09:01.028 "strip_size_kb": 0, 00:09:01.028 "state": "configuring", 00:09:01.028 "raid_level": "raid1", 00:09:01.028 "superblock": true, 00:09:01.028 "num_base_bdevs": 3, 00:09:01.028 "num_base_bdevs_discovered": 2, 00:09:01.028 "num_base_bdevs_operational": 3, 00:09:01.028 "base_bdevs_list": [ 00:09:01.028 { 00:09:01.028 "name": "BaseBdev1", 00:09:01.028 "uuid": "178e76db-ebaa-404f-af9a-47576cd9ab83", 00:09:01.028 "is_configured": true, 00:09:01.028 "data_offset": 2048, 00:09:01.028 "data_size": 63488 00:09:01.028 }, 00:09:01.028 { 00:09:01.028 "name": "BaseBdev2", 00:09:01.028 "uuid": "35cfd449-ba8b-40d8-9a8b-90eddabb0560", 00:09:01.028 "is_configured": true, 00:09:01.028 "data_offset": 2048, 00:09:01.028 "data_size": 63488 00:09:01.028 }, 00:09:01.028 { 00:09:01.028 "name": "BaseBdev3", 00:09:01.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.028 "is_configured": false, 00:09:01.028 "data_offset": 0, 00:09:01.028 "data_size": 0 00:09:01.028 } 00:09:01.028 ] 00:09:01.028 }' 00:09:01.028 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.028 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.288 [2024-09-29 16:47:22.801298] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:01.288 [2024-09-29 16:47:22.801495] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:01.288 [2024-09-29 16:47:22.801520] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:01.288 BaseBdev3 00:09:01.288 [2024-09-29 16:47:22.801831] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:01.288 [2024-09-29 16:47:22.801968] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:01.288 [2024-09-29 16:47:22.801979] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:01.288 [2024-09-29 16:47:22.802107] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.288 [ 00:09:01.288 { 00:09:01.288 "name": "BaseBdev3", 00:09:01.288 "aliases": [ 00:09:01.288 "945052c0-b907-476b-ac1a-64d63566c4a1" 00:09:01.288 ], 00:09:01.288 "product_name": "Malloc disk", 00:09:01.288 "block_size": 512, 00:09:01.288 "num_blocks": 65536, 00:09:01.288 "uuid": "945052c0-b907-476b-ac1a-64d63566c4a1", 00:09:01.288 "assigned_rate_limits": { 00:09:01.288 "rw_ios_per_sec": 0, 00:09:01.288 "rw_mbytes_per_sec": 0, 00:09:01.288 "r_mbytes_per_sec": 0, 00:09:01.288 "w_mbytes_per_sec": 0 00:09:01.288 }, 00:09:01.288 "claimed": true, 00:09:01.288 "claim_type": "exclusive_write", 00:09:01.288 "zoned": false, 00:09:01.288 "supported_io_types": { 00:09:01.288 "read": true, 00:09:01.288 "write": true, 00:09:01.288 "unmap": true, 00:09:01.288 "flush": true, 00:09:01.288 "reset": true, 00:09:01.288 "nvme_admin": false, 00:09:01.288 "nvme_io": false, 00:09:01.288 "nvme_io_md": false, 00:09:01.288 "write_zeroes": true, 00:09:01.288 "zcopy": true, 00:09:01.288 "get_zone_info": false, 00:09:01.288 "zone_management": false, 00:09:01.288 "zone_append": false, 00:09:01.288 "compare": false, 00:09:01.288 "compare_and_write": false, 00:09:01.288 "abort": true, 00:09:01.288 "seek_hole": false, 00:09:01.288 "seek_data": false, 00:09:01.288 "copy": true, 00:09:01.288 "nvme_iov_md": false 00:09:01.288 }, 00:09:01.288 "memory_domains": [ 00:09:01.288 { 00:09:01.288 "dma_device_id": "system", 00:09:01.288 "dma_device_type": 1 00:09:01.288 }, 00:09:01.288 { 00:09:01.288 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.288 "dma_device_type": 2 00:09:01.288 } 00:09:01.288 ], 00:09:01.288 "driver_specific": {} 00:09:01.288 } 00:09:01.288 ] 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.288 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.289 "name": "Existed_Raid", 00:09:01.289 "uuid": "492c9f7b-d766-413e-a0df-af9a8c7a6be4", 00:09:01.289 "strip_size_kb": 0, 00:09:01.289 "state": "online", 00:09:01.289 "raid_level": "raid1", 00:09:01.289 "superblock": true, 00:09:01.289 "num_base_bdevs": 3, 00:09:01.289 "num_base_bdevs_discovered": 3, 00:09:01.289 "num_base_bdevs_operational": 3, 00:09:01.289 "base_bdevs_list": [ 00:09:01.289 { 00:09:01.289 "name": "BaseBdev1", 00:09:01.289 "uuid": "178e76db-ebaa-404f-af9a-47576cd9ab83", 00:09:01.289 "is_configured": true, 00:09:01.289 "data_offset": 2048, 00:09:01.289 "data_size": 63488 00:09:01.289 }, 00:09:01.289 { 00:09:01.289 "name": "BaseBdev2", 00:09:01.289 "uuid": "35cfd449-ba8b-40d8-9a8b-90eddabb0560", 00:09:01.289 "is_configured": true, 00:09:01.289 "data_offset": 2048, 00:09:01.289 "data_size": 63488 00:09:01.289 }, 00:09:01.289 { 00:09:01.289 "name": "BaseBdev3", 00:09:01.289 "uuid": "945052c0-b907-476b-ac1a-64d63566c4a1", 00:09:01.289 "is_configured": true, 00:09:01.289 "data_offset": 2048, 00:09:01.289 "data_size": 63488 00:09:01.289 } 00:09:01.289 ] 00:09:01.289 }' 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.289 16:47:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.548 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.808 [2024-09-29 16:47:23.224925] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:01.808 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.808 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:01.808 "name": "Existed_Raid", 00:09:01.808 "aliases": [ 00:09:01.808 "492c9f7b-d766-413e-a0df-af9a8c7a6be4" 00:09:01.808 ], 00:09:01.808 "product_name": "Raid Volume", 00:09:01.808 "block_size": 512, 00:09:01.808 "num_blocks": 63488, 00:09:01.808 "uuid": "492c9f7b-d766-413e-a0df-af9a8c7a6be4", 00:09:01.808 "assigned_rate_limits": { 00:09:01.808 "rw_ios_per_sec": 0, 00:09:01.808 "rw_mbytes_per_sec": 0, 00:09:01.808 "r_mbytes_per_sec": 0, 00:09:01.808 "w_mbytes_per_sec": 0 00:09:01.808 }, 00:09:01.808 "claimed": false, 00:09:01.808 "zoned": false, 00:09:01.808 "supported_io_types": { 00:09:01.808 "read": true, 00:09:01.808 "write": true, 00:09:01.808 "unmap": false, 00:09:01.808 "flush": false, 00:09:01.808 "reset": true, 00:09:01.808 "nvme_admin": false, 00:09:01.808 "nvme_io": false, 00:09:01.808 "nvme_io_md": false, 00:09:01.808 "write_zeroes": true, 00:09:01.808 "zcopy": false, 00:09:01.808 "get_zone_info": false, 00:09:01.808 "zone_management": false, 00:09:01.808 "zone_append": false, 00:09:01.808 "compare": false, 00:09:01.808 "compare_and_write": false, 00:09:01.808 "abort": false, 00:09:01.808 "seek_hole": false, 00:09:01.808 "seek_data": false, 00:09:01.808 "copy": false, 00:09:01.808 "nvme_iov_md": false 00:09:01.808 }, 00:09:01.808 "memory_domains": [ 00:09:01.808 { 00:09:01.808 "dma_device_id": "system", 00:09:01.808 "dma_device_type": 1 00:09:01.808 }, 00:09:01.808 { 00:09:01.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.808 "dma_device_type": 2 00:09:01.808 }, 00:09:01.808 { 00:09:01.808 "dma_device_id": "system", 00:09:01.808 "dma_device_type": 1 00:09:01.808 }, 00:09:01.808 { 00:09:01.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.808 "dma_device_type": 2 00:09:01.808 }, 00:09:01.808 { 00:09:01.808 "dma_device_id": "system", 00:09:01.808 "dma_device_type": 1 00:09:01.808 }, 00:09:01.808 { 00:09:01.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:01.808 "dma_device_type": 2 00:09:01.808 } 00:09:01.808 ], 00:09:01.808 "driver_specific": { 00:09:01.808 "raid": { 00:09:01.808 "uuid": "492c9f7b-d766-413e-a0df-af9a8c7a6be4", 00:09:01.808 "strip_size_kb": 0, 00:09:01.808 "state": "online", 00:09:01.808 "raid_level": "raid1", 00:09:01.809 "superblock": true, 00:09:01.809 "num_base_bdevs": 3, 00:09:01.809 "num_base_bdevs_discovered": 3, 00:09:01.809 "num_base_bdevs_operational": 3, 00:09:01.809 "base_bdevs_list": [ 00:09:01.809 { 00:09:01.809 "name": "BaseBdev1", 00:09:01.809 "uuid": "178e76db-ebaa-404f-af9a-47576cd9ab83", 00:09:01.809 "is_configured": true, 00:09:01.809 "data_offset": 2048, 00:09:01.809 "data_size": 63488 00:09:01.809 }, 00:09:01.809 { 00:09:01.809 "name": "BaseBdev2", 00:09:01.809 "uuid": "35cfd449-ba8b-40d8-9a8b-90eddabb0560", 00:09:01.809 "is_configured": true, 00:09:01.809 "data_offset": 2048, 00:09:01.809 "data_size": 63488 00:09:01.809 }, 00:09:01.809 { 00:09:01.809 "name": "BaseBdev3", 00:09:01.809 "uuid": "945052c0-b907-476b-ac1a-64d63566c4a1", 00:09:01.809 "is_configured": true, 00:09:01.809 "data_offset": 2048, 00:09:01.809 "data_size": 63488 00:09:01.809 } 00:09:01.809 ] 00:09:01.809 } 00:09:01.809 } 00:09:01.809 }' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:01.809 BaseBdev2 00:09:01.809 BaseBdev3' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:01.809 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.069 [2024-09-29 16:47:23.504172] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.069 "name": "Existed_Raid", 00:09:02.069 "uuid": "492c9f7b-d766-413e-a0df-af9a8c7a6be4", 00:09:02.069 "strip_size_kb": 0, 00:09:02.069 "state": "online", 00:09:02.069 "raid_level": "raid1", 00:09:02.069 "superblock": true, 00:09:02.069 "num_base_bdevs": 3, 00:09:02.069 "num_base_bdevs_discovered": 2, 00:09:02.069 "num_base_bdevs_operational": 2, 00:09:02.069 "base_bdevs_list": [ 00:09:02.069 { 00:09:02.069 "name": null, 00:09:02.069 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.069 "is_configured": false, 00:09:02.069 "data_offset": 0, 00:09:02.069 "data_size": 63488 00:09:02.069 }, 00:09:02.069 { 00:09:02.069 "name": "BaseBdev2", 00:09:02.069 "uuid": "35cfd449-ba8b-40d8-9a8b-90eddabb0560", 00:09:02.069 "is_configured": true, 00:09:02.069 "data_offset": 2048, 00:09:02.069 "data_size": 63488 00:09:02.069 }, 00:09:02.069 { 00:09:02.069 "name": "BaseBdev3", 00:09:02.069 "uuid": "945052c0-b907-476b-ac1a-64d63566c4a1", 00:09:02.069 "is_configured": true, 00:09:02.069 "data_offset": 2048, 00:09:02.069 "data_size": 63488 00:09:02.069 } 00:09:02.069 ] 00:09:02.069 }' 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.069 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.329 [2024-09-29 16:47:23.974459] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.329 16:47:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.590 [2024-09-29 16:47:24.041706] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:02.590 [2024-09-29 16:47:24.041819] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:02.590 [2024-09-29 16:47:24.053461] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:02.590 [2024-09-29 16:47:24.053507] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:02.590 [2024-09-29 16:47:24.053520] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.590 BaseBdev2 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.590 [ 00:09:02.590 { 00:09:02.590 "name": "BaseBdev2", 00:09:02.590 "aliases": [ 00:09:02.590 "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f" 00:09:02.590 ], 00:09:02.590 "product_name": "Malloc disk", 00:09:02.590 "block_size": 512, 00:09:02.590 "num_blocks": 65536, 00:09:02.590 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:02.590 "assigned_rate_limits": { 00:09:02.590 "rw_ios_per_sec": 0, 00:09:02.590 "rw_mbytes_per_sec": 0, 00:09:02.590 "r_mbytes_per_sec": 0, 00:09:02.590 "w_mbytes_per_sec": 0 00:09:02.590 }, 00:09:02.590 "claimed": false, 00:09:02.590 "zoned": false, 00:09:02.590 "supported_io_types": { 00:09:02.590 "read": true, 00:09:02.590 "write": true, 00:09:02.590 "unmap": true, 00:09:02.590 "flush": true, 00:09:02.590 "reset": true, 00:09:02.590 "nvme_admin": false, 00:09:02.590 "nvme_io": false, 00:09:02.590 "nvme_io_md": false, 00:09:02.590 "write_zeroes": true, 00:09:02.590 "zcopy": true, 00:09:02.590 "get_zone_info": false, 00:09:02.590 "zone_management": false, 00:09:02.590 "zone_append": false, 00:09:02.590 "compare": false, 00:09:02.590 "compare_and_write": false, 00:09:02.590 "abort": true, 00:09:02.590 "seek_hole": false, 00:09:02.590 "seek_data": false, 00:09:02.590 "copy": true, 00:09:02.590 "nvme_iov_md": false 00:09:02.590 }, 00:09:02.590 "memory_domains": [ 00:09:02.590 { 00:09:02.590 "dma_device_id": "system", 00:09:02.590 "dma_device_type": 1 00:09:02.590 }, 00:09:02.590 { 00:09:02.590 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.590 "dma_device_type": 2 00:09:02.590 } 00:09:02.590 ], 00:09:02.590 "driver_specific": {} 00:09:02.590 } 00:09:02.590 ] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.590 BaseBdev3 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.590 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 [ 00:09:02.591 { 00:09:02.591 "name": "BaseBdev3", 00:09:02.591 "aliases": [ 00:09:02.591 "0e846258-9fef-45de-9ec4-d9668b79e588" 00:09:02.591 ], 00:09:02.591 "product_name": "Malloc disk", 00:09:02.591 "block_size": 512, 00:09:02.591 "num_blocks": 65536, 00:09:02.591 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:02.591 "assigned_rate_limits": { 00:09:02.591 "rw_ios_per_sec": 0, 00:09:02.591 "rw_mbytes_per_sec": 0, 00:09:02.591 "r_mbytes_per_sec": 0, 00:09:02.591 "w_mbytes_per_sec": 0 00:09:02.591 }, 00:09:02.591 "claimed": false, 00:09:02.591 "zoned": false, 00:09:02.591 "supported_io_types": { 00:09:02.591 "read": true, 00:09:02.591 "write": true, 00:09:02.591 "unmap": true, 00:09:02.591 "flush": true, 00:09:02.591 "reset": true, 00:09:02.591 "nvme_admin": false, 00:09:02.591 "nvme_io": false, 00:09:02.591 "nvme_io_md": false, 00:09:02.591 "write_zeroes": true, 00:09:02.591 "zcopy": true, 00:09:02.591 "get_zone_info": false, 00:09:02.591 "zone_management": false, 00:09:02.591 "zone_append": false, 00:09:02.591 "compare": false, 00:09:02.591 "compare_and_write": false, 00:09:02.591 "abort": true, 00:09:02.591 "seek_hole": false, 00:09:02.591 "seek_data": false, 00:09:02.591 "copy": true, 00:09:02.591 "nvme_iov_md": false 00:09:02.591 }, 00:09:02.591 "memory_domains": [ 00:09:02.591 { 00:09:02.591 "dma_device_id": "system", 00:09:02.591 "dma_device_type": 1 00:09:02.591 }, 00:09:02.591 { 00:09:02.591 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:02.591 "dma_device_type": 2 00:09:02.591 } 00:09:02.591 ], 00:09:02.591 "driver_specific": {} 00:09:02.591 } 00:09:02.591 ] 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 [2024-09-29 16:47:24.201667] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:02.591 [2024-09-29 16:47:24.201729] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:02.591 [2024-09-29 16:47:24.201757] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:02.591 [2024-09-29 16:47:24.203501] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.591 "name": "Existed_Raid", 00:09:02.591 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:02.591 "strip_size_kb": 0, 00:09:02.591 "state": "configuring", 00:09:02.591 "raid_level": "raid1", 00:09:02.591 "superblock": true, 00:09:02.591 "num_base_bdevs": 3, 00:09:02.591 "num_base_bdevs_discovered": 2, 00:09:02.591 "num_base_bdevs_operational": 3, 00:09:02.591 "base_bdevs_list": [ 00:09:02.591 { 00:09:02.591 "name": "BaseBdev1", 00:09:02.591 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.591 "is_configured": false, 00:09:02.591 "data_offset": 0, 00:09:02.591 "data_size": 0 00:09:02.591 }, 00:09:02.591 { 00:09:02.591 "name": "BaseBdev2", 00:09:02.591 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:02.591 "is_configured": true, 00:09:02.591 "data_offset": 2048, 00:09:02.591 "data_size": 63488 00:09:02.591 }, 00:09:02.591 { 00:09:02.591 "name": "BaseBdev3", 00:09:02.591 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:02.591 "is_configured": true, 00:09:02.591 "data_offset": 2048, 00:09:02.591 "data_size": 63488 00:09:02.591 } 00:09:02.591 ] 00:09:02.591 }' 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.591 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.160 [2024-09-29 16:47:24.632898] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.160 "name": "Existed_Raid", 00:09:03.160 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:03.160 "strip_size_kb": 0, 00:09:03.160 "state": "configuring", 00:09:03.160 "raid_level": "raid1", 00:09:03.160 "superblock": true, 00:09:03.160 "num_base_bdevs": 3, 00:09:03.160 "num_base_bdevs_discovered": 1, 00:09:03.160 "num_base_bdevs_operational": 3, 00:09:03.160 "base_bdevs_list": [ 00:09:03.160 { 00:09:03.160 "name": "BaseBdev1", 00:09:03.160 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:03.160 "is_configured": false, 00:09:03.160 "data_offset": 0, 00:09:03.160 "data_size": 0 00:09:03.160 }, 00:09:03.160 { 00:09:03.160 "name": null, 00:09:03.160 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:03.160 "is_configured": false, 00:09:03.160 "data_offset": 0, 00:09:03.160 "data_size": 63488 00:09:03.160 }, 00:09:03.160 { 00:09:03.160 "name": "BaseBdev3", 00:09:03.160 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:03.160 "is_configured": true, 00:09:03.160 "data_offset": 2048, 00:09:03.160 "data_size": 63488 00:09:03.160 } 00:09:03.160 ] 00:09:03.160 }' 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.160 16:47:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.420 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.679 [2024-09-29 16:47:25.095325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:03.679 BaseBdev1 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:03.679 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.680 [ 00:09:03.680 { 00:09:03.680 "name": "BaseBdev1", 00:09:03.680 "aliases": [ 00:09:03.680 "f28cb470-ecc2-438b-b25d-3dd989dd304c" 00:09:03.680 ], 00:09:03.680 "product_name": "Malloc disk", 00:09:03.680 "block_size": 512, 00:09:03.680 "num_blocks": 65536, 00:09:03.680 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:03.680 "assigned_rate_limits": { 00:09:03.680 "rw_ios_per_sec": 0, 00:09:03.680 "rw_mbytes_per_sec": 0, 00:09:03.680 "r_mbytes_per_sec": 0, 00:09:03.680 "w_mbytes_per_sec": 0 00:09:03.680 }, 00:09:03.680 "claimed": true, 00:09:03.680 "claim_type": "exclusive_write", 00:09:03.680 "zoned": false, 00:09:03.680 "supported_io_types": { 00:09:03.680 "read": true, 00:09:03.680 "write": true, 00:09:03.680 "unmap": true, 00:09:03.680 "flush": true, 00:09:03.680 "reset": true, 00:09:03.680 "nvme_admin": false, 00:09:03.680 "nvme_io": false, 00:09:03.680 "nvme_io_md": false, 00:09:03.680 "write_zeroes": true, 00:09:03.680 "zcopy": true, 00:09:03.680 "get_zone_info": false, 00:09:03.680 "zone_management": false, 00:09:03.680 "zone_append": false, 00:09:03.680 "compare": false, 00:09:03.680 "compare_and_write": false, 00:09:03.680 "abort": true, 00:09:03.680 "seek_hole": false, 00:09:03.680 "seek_data": false, 00:09:03.680 "copy": true, 00:09:03.680 "nvme_iov_md": false 00:09:03.680 }, 00:09:03.680 "memory_domains": [ 00:09:03.680 { 00:09:03.680 "dma_device_id": "system", 00:09:03.680 "dma_device_type": 1 00:09:03.680 }, 00:09:03.680 { 00:09:03.680 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.680 "dma_device_type": 2 00:09:03.680 } 00:09:03.680 ], 00:09:03.680 "driver_specific": {} 00:09:03.680 } 00:09:03.680 ] 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.680 "name": "Existed_Raid", 00:09:03.680 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:03.680 "strip_size_kb": 0, 00:09:03.680 "state": "configuring", 00:09:03.680 "raid_level": "raid1", 00:09:03.680 "superblock": true, 00:09:03.680 "num_base_bdevs": 3, 00:09:03.680 "num_base_bdevs_discovered": 2, 00:09:03.680 "num_base_bdevs_operational": 3, 00:09:03.680 "base_bdevs_list": [ 00:09:03.680 { 00:09:03.680 "name": "BaseBdev1", 00:09:03.680 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:03.680 "is_configured": true, 00:09:03.680 "data_offset": 2048, 00:09:03.680 "data_size": 63488 00:09:03.680 }, 00:09:03.680 { 00:09:03.680 "name": null, 00:09:03.680 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:03.680 "is_configured": false, 00:09:03.680 "data_offset": 0, 00:09:03.680 "data_size": 63488 00:09:03.680 }, 00:09:03.680 { 00:09:03.680 "name": "BaseBdev3", 00:09:03.680 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:03.680 "is_configured": true, 00:09:03.680 "data_offset": 2048, 00:09:03.680 "data_size": 63488 00:09:03.680 } 00:09:03.680 ] 00:09:03.680 }' 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.680 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:03.940 [2024-09-29 16:47:25.590581] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.940 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.200 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.200 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.200 "name": "Existed_Raid", 00:09:04.200 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:04.200 "strip_size_kb": 0, 00:09:04.200 "state": "configuring", 00:09:04.200 "raid_level": "raid1", 00:09:04.200 "superblock": true, 00:09:04.200 "num_base_bdevs": 3, 00:09:04.200 "num_base_bdevs_discovered": 1, 00:09:04.200 "num_base_bdevs_operational": 3, 00:09:04.200 "base_bdevs_list": [ 00:09:04.200 { 00:09:04.200 "name": "BaseBdev1", 00:09:04.200 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:04.200 "is_configured": true, 00:09:04.200 "data_offset": 2048, 00:09:04.200 "data_size": 63488 00:09:04.200 }, 00:09:04.200 { 00:09:04.200 "name": null, 00:09:04.200 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:04.200 "is_configured": false, 00:09:04.200 "data_offset": 0, 00:09:04.200 "data_size": 63488 00:09:04.200 }, 00:09:04.200 { 00:09:04.200 "name": null, 00:09:04.200 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:04.200 "is_configured": false, 00:09:04.200 "data_offset": 0, 00:09:04.200 "data_size": 63488 00:09:04.200 } 00:09:04.200 ] 00:09:04.200 }' 00:09:04.200 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.200 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.460 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:04.460 16:47:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.460 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.460 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.460 16:47:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.460 [2024-09-29 16:47:26.009846] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.460 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:04.460 "name": "Existed_Raid", 00:09:04.460 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:04.460 "strip_size_kb": 0, 00:09:04.460 "state": "configuring", 00:09:04.460 "raid_level": "raid1", 00:09:04.460 "superblock": true, 00:09:04.460 "num_base_bdevs": 3, 00:09:04.460 "num_base_bdevs_discovered": 2, 00:09:04.460 "num_base_bdevs_operational": 3, 00:09:04.460 "base_bdevs_list": [ 00:09:04.460 { 00:09:04.460 "name": "BaseBdev1", 00:09:04.460 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:04.460 "is_configured": true, 00:09:04.461 "data_offset": 2048, 00:09:04.461 "data_size": 63488 00:09:04.461 }, 00:09:04.461 { 00:09:04.461 "name": null, 00:09:04.461 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:04.461 "is_configured": false, 00:09:04.461 "data_offset": 0, 00:09:04.461 "data_size": 63488 00:09:04.461 }, 00:09:04.461 { 00:09:04.461 "name": "BaseBdev3", 00:09:04.461 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:04.461 "is_configured": true, 00:09:04.461 "data_offset": 2048, 00:09:04.461 "data_size": 63488 00:09:04.461 } 00:09:04.461 ] 00:09:04.461 }' 00:09:04.461 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:04.461 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.030 [2024-09-29 16:47:26.469083] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.030 "name": "Existed_Raid", 00:09:05.030 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:05.030 "strip_size_kb": 0, 00:09:05.030 "state": "configuring", 00:09:05.030 "raid_level": "raid1", 00:09:05.030 "superblock": true, 00:09:05.030 "num_base_bdevs": 3, 00:09:05.030 "num_base_bdevs_discovered": 1, 00:09:05.030 "num_base_bdevs_operational": 3, 00:09:05.030 "base_bdevs_list": [ 00:09:05.030 { 00:09:05.030 "name": null, 00:09:05.030 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:05.030 "is_configured": false, 00:09:05.030 "data_offset": 0, 00:09:05.030 "data_size": 63488 00:09:05.030 }, 00:09:05.030 { 00:09:05.030 "name": null, 00:09:05.030 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:05.030 "is_configured": false, 00:09:05.030 "data_offset": 0, 00:09:05.030 "data_size": 63488 00:09:05.030 }, 00:09:05.030 { 00:09:05.030 "name": "BaseBdev3", 00:09:05.030 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:05.030 "is_configured": true, 00:09:05.030 "data_offset": 2048, 00:09:05.030 "data_size": 63488 00:09:05.030 } 00:09:05.030 ] 00:09:05.030 }' 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.030 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.290 [2024-09-29 16:47:26.946461] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.290 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.549 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.549 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.549 "name": "Existed_Raid", 00:09:05.549 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:05.549 "strip_size_kb": 0, 00:09:05.549 "state": "configuring", 00:09:05.549 "raid_level": "raid1", 00:09:05.549 "superblock": true, 00:09:05.549 "num_base_bdevs": 3, 00:09:05.549 "num_base_bdevs_discovered": 2, 00:09:05.549 "num_base_bdevs_operational": 3, 00:09:05.549 "base_bdevs_list": [ 00:09:05.549 { 00:09:05.549 "name": null, 00:09:05.549 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:05.549 "is_configured": false, 00:09:05.549 "data_offset": 0, 00:09:05.549 "data_size": 63488 00:09:05.550 }, 00:09:05.550 { 00:09:05.550 "name": "BaseBdev2", 00:09:05.550 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:05.550 "is_configured": true, 00:09:05.550 "data_offset": 2048, 00:09:05.550 "data_size": 63488 00:09:05.550 }, 00:09:05.550 { 00:09:05.550 "name": "BaseBdev3", 00:09:05.550 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:05.550 "is_configured": true, 00:09:05.550 "data_offset": 2048, 00:09:05.550 "data_size": 63488 00:09:05.550 } 00:09:05.550 ] 00:09:05.550 }' 00:09:05.550 16:47:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.550 16:47:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u f28cb470-ecc2-438b-b25d-3dd989dd304c 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.810 [2024-09-29 16:47:27.416591] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:05.810 [2024-09-29 16:47:27.416785] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:05.810 [2024-09-29 16:47:27.416805] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:05.810 NewBaseBdev 00:09:05.810 [2024-09-29 16:47:27.417080] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:05.810 [2024-09-29 16:47:27.417198] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:05.810 [2024-09-29 16:47:27.417212] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:05.810 [2024-09-29 16:47:27.417310] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.810 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.810 [ 00:09:05.810 { 00:09:05.810 "name": "NewBaseBdev", 00:09:05.810 "aliases": [ 00:09:05.810 "f28cb470-ecc2-438b-b25d-3dd989dd304c" 00:09:05.810 ], 00:09:05.810 "product_name": "Malloc disk", 00:09:05.810 "block_size": 512, 00:09:05.810 "num_blocks": 65536, 00:09:05.810 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:05.810 "assigned_rate_limits": { 00:09:05.810 "rw_ios_per_sec": 0, 00:09:05.810 "rw_mbytes_per_sec": 0, 00:09:05.810 "r_mbytes_per_sec": 0, 00:09:05.810 "w_mbytes_per_sec": 0 00:09:05.810 }, 00:09:05.810 "claimed": true, 00:09:05.810 "claim_type": "exclusive_write", 00:09:05.810 "zoned": false, 00:09:05.810 "supported_io_types": { 00:09:05.810 "read": true, 00:09:05.810 "write": true, 00:09:05.810 "unmap": true, 00:09:05.810 "flush": true, 00:09:05.810 "reset": true, 00:09:05.810 "nvme_admin": false, 00:09:05.810 "nvme_io": false, 00:09:05.810 "nvme_io_md": false, 00:09:05.810 "write_zeroes": true, 00:09:05.810 "zcopy": true, 00:09:05.810 "get_zone_info": false, 00:09:05.810 "zone_management": false, 00:09:05.810 "zone_append": false, 00:09:05.810 "compare": false, 00:09:05.811 "compare_and_write": false, 00:09:05.811 "abort": true, 00:09:05.811 "seek_hole": false, 00:09:05.811 "seek_data": false, 00:09:05.811 "copy": true, 00:09:05.811 "nvme_iov_md": false 00:09:05.811 }, 00:09:05.811 "memory_domains": [ 00:09:05.811 { 00:09:05.811 "dma_device_id": "system", 00:09:05.811 "dma_device_type": 1 00:09:05.811 }, 00:09:05.811 { 00:09:05.811 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.811 "dma_device_type": 2 00:09:05.811 } 00:09:05.811 ], 00:09:05.811 "driver_specific": {} 00:09:05.811 } 00:09:05.811 ] 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.811 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.071 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.071 "name": "Existed_Raid", 00:09:06.071 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:06.071 "strip_size_kb": 0, 00:09:06.071 "state": "online", 00:09:06.071 "raid_level": "raid1", 00:09:06.071 "superblock": true, 00:09:06.071 "num_base_bdevs": 3, 00:09:06.071 "num_base_bdevs_discovered": 3, 00:09:06.071 "num_base_bdevs_operational": 3, 00:09:06.071 "base_bdevs_list": [ 00:09:06.071 { 00:09:06.071 "name": "NewBaseBdev", 00:09:06.071 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:06.071 "is_configured": true, 00:09:06.071 "data_offset": 2048, 00:09:06.071 "data_size": 63488 00:09:06.071 }, 00:09:06.071 { 00:09:06.071 "name": "BaseBdev2", 00:09:06.071 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:06.071 "is_configured": true, 00:09:06.071 "data_offset": 2048, 00:09:06.071 "data_size": 63488 00:09:06.071 }, 00:09:06.071 { 00:09:06.071 "name": "BaseBdev3", 00:09:06.071 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:06.071 "is_configured": true, 00:09:06.071 "data_offset": 2048, 00:09:06.071 "data_size": 63488 00:09:06.071 } 00:09:06.071 ] 00:09:06.071 }' 00:09:06.071 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.071 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.331 [2024-09-29 16:47:27.872113] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:06.331 "name": "Existed_Raid", 00:09:06.331 "aliases": [ 00:09:06.331 "109408dc-d78d-4de2-96ff-5265f803ad7c" 00:09:06.331 ], 00:09:06.331 "product_name": "Raid Volume", 00:09:06.331 "block_size": 512, 00:09:06.331 "num_blocks": 63488, 00:09:06.331 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:06.331 "assigned_rate_limits": { 00:09:06.331 "rw_ios_per_sec": 0, 00:09:06.331 "rw_mbytes_per_sec": 0, 00:09:06.331 "r_mbytes_per_sec": 0, 00:09:06.331 "w_mbytes_per_sec": 0 00:09:06.331 }, 00:09:06.331 "claimed": false, 00:09:06.331 "zoned": false, 00:09:06.331 "supported_io_types": { 00:09:06.331 "read": true, 00:09:06.331 "write": true, 00:09:06.331 "unmap": false, 00:09:06.331 "flush": false, 00:09:06.331 "reset": true, 00:09:06.331 "nvme_admin": false, 00:09:06.331 "nvme_io": false, 00:09:06.331 "nvme_io_md": false, 00:09:06.331 "write_zeroes": true, 00:09:06.331 "zcopy": false, 00:09:06.331 "get_zone_info": false, 00:09:06.331 "zone_management": false, 00:09:06.331 "zone_append": false, 00:09:06.331 "compare": false, 00:09:06.331 "compare_and_write": false, 00:09:06.331 "abort": false, 00:09:06.331 "seek_hole": false, 00:09:06.331 "seek_data": false, 00:09:06.331 "copy": false, 00:09:06.331 "nvme_iov_md": false 00:09:06.331 }, 00:09:06.331 "memory_domains": [ 00:09:06.331 { 00:09:06.331 "dma_device_id": "system", 00:09:06.331 "dma_device_type": 1 00:09:06.331 }, 00:09:06.331 { 00:09:06.331 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.331 "dma_device_type": 2 00:09:06.331 }, 00:09:06.331 { 00:09:06.331 "dma_device_id": "system", 00:09:06.331 "dma_device_type": 1 00:09:06.331 }, 00:09:06.331 { 00:09:06.331 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.331 "dma_device_type": 2 00:09:06.331 }, 00:09:06.331 { 00:09:06.331 "dma_device_id": "system", 00:09:06.331 "dma_device_type": 1 00:09:06.331 }, 00:09:06.331 { 00:09:06.331 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.331 "dma_device_type": 2 00:09:06.331 } 00:09:06.331 ], 00:09:06.331 "driver_specific": { 00:09:06.331 "raid": { 00:09:06.331 "uuid": "109408dc-d78d-4de2-96ff-5265f803ad7c", 00:09:06.331 "strip_size_kb": 0, 00:09:06.331 "state": "online", 00:09:06.331 "raid_level": "raid1", 00:09:06.331 "superblock": true, 00:09:06.331 "num_base_bdevs": 3, 00:09:06.331 "num_base_bdevs_discovered": 3, 00:09:06.331 "num_base_bdevs_operational": 3, 00:09:06.331 "base_bdevs_list": [ 00:09:06.331 { 00:09:06.331 "name": "NewBaseBdev", 00:09:06.331 "uuid": "f28cb470-ecc2-438b-b25d-3dd989dd304c", 00:09:06.331 "is_configured": true, 00:09:06.331 "data_offset": 2048, 00:09:06.331 "data_size": 63488 00:09:06.331 }, 00:09:06.331 { 00:09:06.331 "name": "BaseBdev2", 00:09:06.331 "uuid": "ddd3e9e0-7fb1-4d2d-932d-235ec3878a2f", 00:09:06.331 "is_configured": true, 00:09:06.331 "data_offset": 2048, 00:09:06.331 "data_size": 63488 00:09:06.331 }, 00:09:06.331 { 00:09:06.331 "name": "BaseBdev3", 00:09:06.331 "uuid": "0e846258-9fef-45de-9ec4-d9668b79e588", 00:09:06.331 "is_configured": true, 00:09:06.331 "data_offset": 2048, 00:09:06.331 "data_size": 63488 00:09:06.331 } 00:09:06.331 ] 00:09:06.331 } 00:09:06.331 } 00:09:06.331 }' 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:06.331 BaseBdev2 00:09:06.331 BaseBdev3' 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.331 16:47:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.591 [2024-09-29 16:47:28.099444] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:06.591 [2024-09-29 16:47:28.099472] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:06.591 [2024-09-29 16:47:28.099531] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:06.591 [2024-09-29 16:47:28.099778] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:06.591 [2024-09-29 16:47:28.099806] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78757 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 78757 ']' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 78757 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78757 00:09:06.591 killing process with pid 78757 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78757' 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 78757 00:09:06.591 [2024-09-29 16:47:28.144124] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:06.591 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 78757 00:09:06.591 [2024-09-29 16:47:28.174426] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:06.851 16:47:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:06.851 00:09:06.851 real 0m8.352s 00:09:06.851 user 0m14.262s 00:09:06.851 sys 0m1.696s 00:09:06.851 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:06.851 16:47:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.851 ************************************ 00:09:06.851 END TEST raid_state_function_test_sb 00:09:06.851 ************************************ 00:09:06.851 16:47:28 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:09:06.851 16:47:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:09:06.851 16:47:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:06.851 16:47:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:06.851 ************************************ 00:09:06.851 START TEST raid_superblock_test 00:09:06.851 ************************************ 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 3 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79361 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79361 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 79361 ']' 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:06.851 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:06.851 16:47:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.111 [2024-09-29 16:47:28.558734] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:07.111 [2024-09-29 16:47:28.558869] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79361 ] 00:09:07.111 [2024-09-29 16:47:28.702822] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:07.111 [2024-09-29 16:47:28.748417] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:07.370 [2024-09-29 16:47:28.790710] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.370 [2024-09-29 16:47:28.790759] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.938 malloc1 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.938 [2024-09-29 16:47:29.396759] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:07.938 [2024-09-29 16:47:29.396832] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:07.938 [2024-09-29 16:47:29.396851] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:07.938 [2024-09-29 16:47:29.396865] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:07.938 [2024-09-29 16:47:29.398945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:07.938 [2024-09-29 16:47:29.398984] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:07.938 pt1 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.938 malloc2 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.938 [2024-09-29 16:47:29.436813] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:07.938 [2024-09-29 16:47:29.436864] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:07.938 [2024-09-29 16:47:29.436881] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:07.938 [2024-09-29 16:47:29.436893] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:07.938 [2024-09-29 16:47:29.439127] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:07.938 [2024-09-29 16:47:29.439164] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:07.938 pt2 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:07.938 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.939 malloc3 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.939 [2024-09-29 16:47:29.469557] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:07.939 [2024-09-29 16:47:29.469607] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:07.939 [2024-09-29 16:47:29.469626] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:07.939 [2024-09-29 16:47:29.469636] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:07.939 [2024-09-29 16:47:29.471832] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:07.939 [2024-09-29 16:47:29.471869] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:07.939 pt3 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.939 [2024-09-29 16:47:29.477613] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:07.939 [2024-09-29 16:47:29.479538] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:07.939 [2024-09-29 16:47:29.479597] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:07.939 [2024-09-29 16:47:29.479765] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:07.939 [2024-09-29 16:47:29.479785] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:07.939 [2024-09-29 16:47:29.480073] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:07.939 [2024-09-29 16:47:29.480212] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:07.939 [2024-09-29 16:47:29.480243] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:07.939 [2024-09-29 16:47:29.480392] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:07.939 "name": "raid_bdev1", 00:09:07.939 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:07.939 "strip_size_kb": 0, 00:09:07.939 "state": "online", 00:09:07.939 "raid_level": "raid1", 00:09:07.939 "superblock": true, 00:09:07.939 "num_base_bdevs": 3, 00:09:07.939 "num_base_bdevs_discovered": 3, 00:09:07.939 "num_base_bdevs_operational": 3, 00:09:07.939 "base_bdevs_list": [ 00:09:07.939 { 00:09:07.939 "name": "pt1", 00:09:07.939 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:07.939 "is_configured": true, 00:09:07.939 "data_offset": 2048, 00:09:07.939 "data_size": 63488 00:09:07.939 }, 00:09:07.939 { 00:09:07.939 "name": "pt2", 00:09:07.939 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:07.939 "is_configured": true, 00:09:07.939 "data_offset": 2048, 00:09:07.939 "data_size": 63488 00:09:07.939 }, 00:09:07.939 { 00:09:07.939 "name": "pt3", 00:09:07.939 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:07.939 "is_configured": true, 00:09:07.939 "data_offset": 2048, 00:09:07.939 "data_size": 63488 00:09:07.939 } 00:09:07.939 ] 00:09:07.939 }' 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:07.939 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:08.509 [2024-09-29 16:47:29.889119] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:08.509 "name": "raid_bdev1", 00:09:08.509 "aliases": [ 00:09:08.509 "de86c596-e504-44c3-9c8c-d34e57b67985" 00:09:08.509 ], 00:09:08.509 "product_name": "Raid Volume", 00:09:08.509 "block_size": 512, 00:09:08.509 "num_blocks": 63488, 00:09:08.509 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:08.509 "assigned_rate_limits": { 00:09:08.509 "rw_ios_per_sec": 0, 00:09:08.509 "rw_mbytes_per_sec": 0, 00:09:08.509 "r_mbytes_per_sec": 0, 00:09:08.509 "w_mbytes_per_sec": 0 00:09:08.509 }, 00:09:08.509 "claimed": false, 00:09:08.509 "zoned": false, 00:09:08.509 "supported_io_types": { 00:09:08.509 "read": true, 00:09:08.509 "write": true, 00:09:08.509 "unmap": false, 00:09:08.509 "flush": false, 00:09:08.509 "reset": true, 00:09:08.509 "nvme_admin": false, 00:09:08.509 "nvme_io": false, 00:09:08.509 "nvme_io_md": false, 00:09:08.509 "write_zeroes": true, 00:09:08.509 "zcopy": false, 00:09:08.509 "get_zone_info": false, 00:09:08.509 "zone_management": false, 00:09:08.509 "zone_append": false, 00:09:08.509 "compare": false, 00:09:08.509 "compare_and_write": false, 00:09:08.509 "abort": false, 00:09:08.509 "seek_hole": false, 00:09:08.509 "seek_data": false, 00:09:08.509 "copy": false, 00:09:08.509 "nvme_iov_md": false 00:09:08.509 }, 00:09:08.509 "memory_domains": [ 00:09:08.509 { 00:09:08.509 "dma_device_id": "system", 00:09:08.509 "dma_device_type": 1 00:09:08.509 }, 00:09:08.509 { 00:09:08.509 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:08.509 "dma_device_type": 2 00:09:08.509 }, 00:09:08.509 { 00:09:08.509 "dma_device_id": "system", 00:09:08.509 "dma_device_type": 1 00:09:08.509 }, 00:09:08.509 { 00:09:08.509 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:08.509 "dma_device_type": 2 00:09:08.509 }, 00:09:08.509 { 00:09:08.509 "dma_device_id": "system", 00:09:08.509 "dma_device_type": 1 00:09:08.509 }, 00:09:08.509 { 00:09:08.509 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:08.509 "dma_device_type": 2 00:09:08.509 } 00:09:08.509 ], 00:09:08.509 "driver_specific": { 00:09:08.509 "raid": { 00:09:08.509 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:08.509 "strip_size_kb": 0, 00:09:08.509 "state": "online", 00:09:08.509 "raid_level": "raid1", 00:09:08.509 "superblock": true, 00:09:08.509 "num_base_bdevs": 3, 00:09:08.509 "num_base_bdevs_discovered": 3, 00:09:08.509 "num_base_bdevs_operational": 3, 00:09:08.509 "base_bdevs_list": [ 00:09:08.509 { 00:09:08.509 "name": "pt1", 00:09:08.509 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:08.509 "is_configured": true, 00:09:08.509 "data_offset": 2048, 00:09:08.509 "data_size": 63488 00:09:08.509 }, 00:09:08.509 { 00:09:08.509 "name": "pt2", 00:09:08.509 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:08.509 "is_configured": true, 00:09:08.509 "data_offset": 2048, 00:09:08.509 "data_size": 63488 00:09:08.509 }, 00:09:08.509 { 00:09:08.509 "name": "pt3", 00:09:08.509 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:08.509 "is_configured": true, 00:09:08.509 "data_offset": 2048, 00:09:08.509 "data_size": 63488 00:09:08.509 } 00:09:08.509 ] 00:09:08.509 } 00:09:08.509 } 00:09:08.509 }' 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:08.509 pt2 00:09:08.509 pt3' 00:09:08.509 16:47:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:08.509 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:08.509 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:08.509 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:08.509 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:08.509 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.509 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.509 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.510 [2024-09-29 16:47:30.140637] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=de86c596-e504-44c3-9c8c-d34e57b67985 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z de86c596-e504-44c3-9c8c-d34e57b67985 ']' 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.510 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 [2024-09-29 16:47:30.184328] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:08.770 [2024-09-29 16:47:30.184354] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:08.770 [2024-09-29 16:47:30.184455] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:08.770 [2024-09-29 16:47:30.184532] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:08.770 [2024-09-29 16:47:30.184547] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 [2024-09-29 16:47:30.324115] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:08.770 [2024-09-29 16:47:30.325930] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:08.770 [2024-09-29 16:47:30.325984] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:08.770 [2024-09-29 16:47:30.326033] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:08.770 [2024-09-29 16:47:30.326070] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:08.770 [2024-09-29 16:47:30.326087] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:08.770 [2024-09-29 16:47:30.326099] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:08.770 [2024-09-29 16:47:30.326108] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:08.770 request: 00:09:08.770 { 00:09:08.770 "name": "raid_bdev1", 00:09:08.770 "raid_level": "raid1", 00:09:08.770 "base_bdevs": [ 00:09:08.770 "malloc1", 00:09:08.770 "malloc2", 00:09:08.770 "malloc3" 00:09:08.770 ], 00:09:08.770 "superblock": false, 00:09:08.770 "method": "bdev_raid_create", 00:09:08.770 "req_id": 1 00:09:08.770 } 00:09:08.770 Got JSON-RPC error response 00:09:08.770 response: 00:09:08.770 { 00:09:08.770 "code": -17, 00:09:08.770 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:08.770 } 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:08.770 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.771 [2024-09-29 16:47:30.387968] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:08.771 [2024-09-29 16:47:30.388015] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:08.771 [2024-09-29 16:47:30.388030] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:08.771 [2024-09-29 16:47:30.388040] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:08.771 [2024-09-29 16:47:30.390131] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:08.771 [2024-09-29 16:47:30.390168] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:08.771 [2024-09-29 16:47:30.390225] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:08.771 [2024-09-29 16:47:30.390255] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:08.771 pt1 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:08.771 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.030 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.030 "name": "raid_bdev1", 00:09:09.030 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:09.030 "strip_size_kb": 0, 00:09:09.030 "state": "configuring", 00:09:09.031 "raid_level": "raid1", 00:09:09.031 "superblock": true, 00:09:09.031 "num_base_bdevs": 3, 00:09:09.031 "num_base_bdevs_discovered": 1, 00:09:09.031 "num_base_bdevs_operational": 3, 00:09:09.031 "base_bdevs_list": [ 00:09:09.031 { 00:09:09.031 "name": "pt1", 00:09:09.031 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:09.031 "is_configured": true, 00:09:09.031 "data_offset": 2048, 00:09:09.031 "data_size": 63488 00:09:09.031 }, 00:09:09.031 { 00:09:09.031 "name": null, 00:09:09.031 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:09.031 "is_configured": false, 00:09:09.031 "data_offset": 2048, 00:09:09.031 "data_size": 63488 00:09:09.031 }, 00:09:09.031 { 00:09:09.031 "name": null, 00:09:09.031 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:09.031 "is_configured": false, 00:09:09.031 "data_offset": 2048, 00:09:09.031 "data_size": 63488 00:09:09.031 } 00:09:09.031 ] 00:09:09.031 }' 00:09:09.031 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.031 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.290 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:09.290 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:09.290 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.290 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.290 [2024-09-29 16:47:30.827250] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:09.290 [2024-09-29 16:47:30.827314] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:09.290 [2024-09-29 16:47:30.827334] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:09.291 [2024-09-29 16:47:30.827347] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:09.291 [2024-09-29 16:47:30.827754] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:09.291 [2024-09-29 16:47:30.827783] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:09.291 [2024-09-29 16:47:30.827850] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:09.291 [2024-09-29 16:47:30.827872] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:09.291 pt2 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.291 [2024-09-29 16:47:30.839225] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.291 "name": "raid_bdev1", 00:09:09.291 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:09.291 "strip_size_kb": 0, 00:09:09.291 "state": "configuring", 00:09:09.291 "raid_level": "raid1", 00:09:09.291 "superblock": true, 00:09:09.291 "num_base_bdevs": 3, 00:09:09.291 "num_base_bdevs_discovered": 1, 00:09:09.291 "num_base_bdevs_operational": 3, 00:09:09.291 "base_bdevs_list": [ 00:09:09.291 { 00:09:09.291 "name": "pt1", 00:09:09.291 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:09.291 "is_configured": true, 00:09:09.291 "data_offset": 2048, 00:09:09.291 "data_size": 63488 00:09:09.291 }, 00:09:09.291 { 00:09:09.291 "name": null, 00:09:09.291 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:09.291 "is_configured": false, 00:09:09.291 "data_offset": 0, 00:09:09.291 "data_size": 63488 00:09:09.291 }, 00:09:09.291 { 00:09:09.291 "name": null, 00:09:09.291 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:09.291 "is_configured": false, 00:09:09.291 "data_offset": 2048, 00:09:09.291 "data_size": 63488 00:09:09.291 } 00:09:09.291 ] 00:09:09.291 }' 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.291 16:47:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.860 [2024-09-29 16:47:31.310409] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:09.860 [2024-09-29 16:47:31.310456] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:09.860 [2024-09-29 16:47:31.310476] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:09.860 [2024-09-29 16:47:31.310503] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:09.860 [2024-09-29 16:47:31.310913] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:09.860 [2024-09-29 16:47:31.310939] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:09.860 [2024-09-29 16:47:31.311008] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:09.860 [2024-09-29 16:47:31.311046] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:09.860 pt2 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.860 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.860 [2024-09-29 16:47:31.322375] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:09.860 [2024-09-29 16:47:31.322433] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:09.860 [2024-09-29 16:47:31.322450] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:09.860 [2024-09-29 16:47:31.322463] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:09.860 [2024-09-29 16:47:31.322798] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:09.860 [2024-09-29 16:47:31.322821] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:09.860 [2024-09-29 16:47:31.322877] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:09.861 [2024-09-29 16:47:31.322916] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:09.861 [2024-09-29 16:47:31.323062] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:09.861 [2024-09-29 16:47:31.323076] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:09.861 [2024-09-29 16:47:31.323300] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:09.861 [2024-09-29 16:47:31.323424] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:09.861 [2024-09-29 16:47:31.323441] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:09.861 [2024-09-29 16:47:31.323545] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:09.861 pt3 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.861 "name": "raid_bdev1", 00:09:09.861 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:09.861 "strip_size_kb": 0, 00:09:09.861 "state": "online", 00:09:09.861 "raid_level": "raid1", 00:09:09.861 "superblock": true, 00:09:09.861 "num_base_bdevs": 3, 00:09:09.861 "num_base_bdevs_discovered": 3, 00:09:09.861 "num_base_bdevs_operational": 3, 00:09:09.861 "base_bdevs_list": [ 00:09:09.861 { 00:09:09.861 "name": "pt1", 00:09:09.861 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:09.861 "is_configured": true, 00:09:09.861 "data_offset": 2048, 00:09:09.861 "data_size": 63488 00:09:09.861 }, 00:09:09.861 { 00:09:09.861 "name": "pt2", 00:09:09.861 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:09.861 "is_configured": true, 00:09:09.861 "data_offset": 2048, 00:09:09.861 "data_size": 63488 00:09:09.861 }, 00:09:09.861 { 00:09:09.861 "name": "pt3", 00:09:09.861 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:09.861 "is_configured": true, 00:09:09.861 "data_offset": 2048, 00:09:09.861 "data_size": 63488 00:09:09.861 } 00:09:09.861 ] 00:09:09.861 }' 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.861 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:10.120 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.121 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.121 [2024-09-29 16:47:31.785974] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:10.379 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.379 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:10.379 "name": "raid_bdev1", 00:09:10.379 "aliases": [ 00:09:10.379 "de86c596-e504-44c3-9c8c-d34e57b67985" 00:09:10.379 ], 00:09:10.379 "product_name": "Raid Volume", 00:09:10.379 "block_size": 512, 00:09:10.379 "num_blocks": 63488, 00:09:10.379 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:10.379 "assigned_rate_limits": { 00:09:10.379 "rw_ios_per_sec": 0, 00:09:10.379 "rw_mbytes_per_sec": 0, 00:09:10.379 "r_mbytes_per_sec": 0, 00:09:10.379 "w_mbytes_per_sec": 0 00:09:10.379 }, 00:09:10.379 "claimed": false, 00:09:10.379 "zoned": false, 00:09:10.379 "supported_io_types": { 00:09:10.379 "read": true, 00:09:10.379 "write": true, 00:09:10.379 "unmap": false, 00:09:10.379 "flush": false, 00:09:10.379 "reset": true, 00:09:10.379 "nvme_admin": false, 00:09:10.379 "nvme_io": false, 00:09:10.379 "nvme_io_md": false, 00:09:10.379 "write_zeroes": true, 00:09:10.379 "zcopy": false, 00:09:10.379 "get_zone_info": false, 00:09:10.379 "zone_management": false, 00:09:10.379 "zone_append": false, 00:09:10.379 "compare": false, 00:09:10.379 "compare_and_write": false, 00:09:10.379 "abort": false, 00:09:10.379 "seek_hole": false, 00:09:10.379 "seek_data": false, 00:09:10.379 "copy": false, 00:09:10.379 "nvme_iov_md": false 00:09:10.379 }, 00:09:10.379 "memory_domains": [ 00:09:10.379 { 00:09:10.379 "dma_device_id": "system", 00:09:10.379 "dma_device_type": 1 00:09:10.379 }, 00:09:10.379 { 00:09:10.379 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.379 "dma_device_type": 2 00:09:10.379 }, 00:09:10.379 { 00:09:10.379 "dma_device_id": "system", 00:09:10.379 "dma_device_type": 1 00:09:10.379 }, 00:09:10.379 { 00:09:10.379 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.379 "dma_device_type": 2 00:09:10.379 }, 00:09:10.379 { 00:09:10.379 "dma_device_id": "system", 00:09:10.379 "dma_device_type": 1 00:09:10.379 }, 00:09:10.379 { 00:09:10.379 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:10.379 "dma_device_type": 2 00:09:10.379 } 00:09:10.379 ], 00:09:10.379 "driver_specific": { 00:09:10.379 "raid": { 00:09:10.379 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:10.379 "strip_size_kb": 0, 00:09:10.379 "state": "online", 00:09:10.379 "raid_level": "raid1", 00:09:10.379 "superblock": true, 00:09:10.379 "num_base_bdevs": 3, 00:09:10.379 "num_base_bdevs_discovered": 3, 00:09:10.379 "num_base_bdevs_operational": 3, 00:09:10.379 "base_bdevs_list": [ 00:09:10.379 { 00:09:10.379 "name": "pt1", 00:09:10.379 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:10.380 "is_configured": true, 00:09:10.380 "data_offset": 2048, 00:09:10.380 "data_size": 63488 00:09:10.380 }, 00:09:10.380 { 00:09:10.380 "name": "pt2", 00:09:10.380 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:10.380 "is_configured": true, 00:09:10.380 "data_offset": 2048, 00:09:10.380 "data_size": 63488 00:09:10.380 }, 00:09:10.380 { 00:09:10.380 "name": "pt3", 00:09:10.380 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:10.380 "is_configured": true, 00:09:10.380 "data_offset": 2048, 00:09:10.380 "data_size": 63488 00:09:10.380 } 00:09:10.380 ] 00:09:10.380 } 00:09:10.380 } 00:09:10.380 }' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:10.380 pt2 00:09:10.380 pt3' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.380 16:47:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.380 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.380 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:10.380 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:10.380 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:10.380 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:10.380 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.380 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.380 [2024-09-29 16:47:32.045427] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' de86c596-e504-44c3-9c8c-d34e57b67985 '!=' de86c596-e504-44c3-9c8c-d34e57b67985 ']' 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.639 [2024-09-29 16:47:32.085172] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.639 "name": "raid_bdev1", 00:09:10.639 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:10.639 "strip_size_kb": 0, 00:09:10.639 "state": "online", 00:09:10.639 "raid_level": "raid1", 00:09:10.639 "superblock": true, 00:09:10.639 "num_base_bdevs": 3, 00:09:10.639 "num_base_bdevs_discovered": 2, 00:09:10.639 "num_base_bdevs_operational": 2, 00:09:10.639 "base_bdevs_list": [ 00:09:10.639 { 00:09:10.639 "name": null, 00:09:10.639 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:10.639 "is_configured": false, 00:09:10.639 "data_offset": 0, 00:09:10.639 "data_size": 63488 00:09:10.639 }, 00:09:10.639 { 00:09:10.639 "name": "pt2", 00:09:10.639 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:10.639 "is_configured": true, 00:09:10.639 "data_offset": 2048, 00:09:10.639 "data_size": 63488 00:09:10.639 }, 00:09:10.639 { 00:09:10.639 "name": "pt3", 00:09:10.639 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:10.639 "is_configured": true, 00:09:10.639 "data_offset": 2048, 00:09:10.639 "data_size": 63488 00:09:10.639 } 00:09:10.639 ] 00:09:10.639 }' 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.639 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.905 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:10.905 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.905 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.905 [2024-09-29 16:47:32.524400] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:10.905 [2024-09-29 16:47:32.524437] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:10.905 [2024-09-29 16:47:32.524501] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:10.906 [2024-09-29 16:47:32.524569] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:10.906 [2024-09-29 16:47:32.524593] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.906 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.165 [2024-09-29 16:47:32.588287] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:11.165 [2024-09-29 16:47:32.588342] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.165 [2024-09-29 16:47:32.588362] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:09:11.165 [2024-09-29 16:47:32.588370] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.165 [2024-09-29 16:47:32.590457] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.165 [2024-09-29 16:47:32.590493] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:11.165 [2024-09-29 16:47:32.590557] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:11.165 [2024-09-29 16:47:32.590587] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:11.165 pt2 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.165 "name": "raid_bdev1", 00:09:11.165 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:11.165 "strip_size_kb": 0, 00:09:11.165 "state": "configuring", 00:09:11.165 "raid_level": "raid1", 00:09:11.165 "superblock": true, 00:09:11.165 "num_base_bdevs": 3, 00:09:11.165 "num_base_bdevs_discovered": 1, 00:09:11.165 "num_base_bdevs_operational": 2, 00:09:11.165 "base_bdevs_list": [ 00:09:11.165 { 00:09:11.165 "name": null, 00:09:11.165 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.165 "is_configured": false, 00:09:11.165 "data_offset": 2048, 00:09:11.165 "data_size": 63488 00:09:11.165 }, 00:09:11.165 { 00:09:11.165 "name": "pt2", 00:09:11.165 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:11.165 "is_configured": true, 00:09:11.165 "data_offset": 2048, 00:09:11.165 "data_size": 63488 00:09:11.165 }, 00:09:11.165 { 00:09:11.165 "name": null, 00:09:11.165 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:11.165 "is_configured": false, 00:09:11.165 "data_offset": 2048, 00:09:11.165 "data_size": 63488 00:09:11.165 } 00:09:11.165 ] 00:09:11.165 }' 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.165 16:47:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.425 [2024-09-29 16:47:33.035525] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:11.425 [2024-09-29 16:47:33.035583] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.425 [2024-09-29 16:47:33.035604] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:11.425 [2024-09-29 16:47:33.035613] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.425 [2024-09-29 16:47:33.036060] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.425 [2024-09-29 16:47:33.036087] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:11.425 [2024-09-29 16:47:33.036171] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:11.425 [2024-09-29 16:47:33.036206] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:11.425 [2024-09-29 16:47:33.036341] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:11.425 [2024-09-29 16:47:33.036355] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:11.425 [2024-09-29 16:47:33.036596] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:11.425 [2024-09-29 16:47:33.036741] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:11.425 [2024-09-29 16:47:33.036761] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:11.425 [2024-09-29 16:47:33.036873] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:11.425 pt3 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.425 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.426 "name": "raid_bdev1", 00:09:11.426 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:11.426 "strip_size_kb": 0, 00:09:11.426 "state": "online", 00:09:11.426 "raid_level": "raid1", 00:09:11.426 "superblock": true, 00:09:11.426 "num_base_bdevs": 3, 00:09:11.426 "num_base_bdevs_discovered": 2, 00:09:11.426 "num_base_bdevs_operational": 2, 00:09:11.426 "base_bdevs_list": [ 00:09:11.426 { 00:09:11.426 "name": null, 00:09:11.426 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.426 "is_configured": false, 00:09:11.426 "data_offset": 2048, 00:09:11.426 "data_size": 63488 00:09:11.426 }, 00:09:11.426 { 00:09:11.426 "name": "pt2", 00:09:11.426 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:11.426 "is_configured": true, 00:09:11.426 "data_offset": 2048, 00:09:11.426 "data_size": 63488 00:09:11.426 }, 00:09:11.426 { 00:09:11.426 "name": "pt3", 00:09:11.426 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:11.426 "is_configured": true, 00:09:11.426 "data_offset": 2048, 00:09:11.426 "data_size": 63488 00:09:11.426 } 00:09:11.426 ] 00:09:11.426 }' 00:09:11.426 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.426 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.994 [2024-09-29 16:47:33.442818] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:11.994 [2024-09-29 16:47:33.442844] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:11.994 [2024-09-29 16:47:33.442921] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:11.994 [2024-09-29 16:47:33.442971] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:11.994 [2024-09-29 16:47:33.442981] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:09:11.994 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.995 [2024-09-29 16:47:33.506699] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:11.995 [2024-09-29 16:47:33.506756] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:11.995 [2024-09-29 16:47:33.506770] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:11.995 [2024-09-29 16:47:33.506780] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:11.995 [2024-09-29 16:47:33.508813] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:11.995 [2024-09-29 16:47:33.508849] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:11.995 [2024-09-29 16:47:33.508923] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:11.995 [2024-09-29 16:47:33.508966] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:11.995 [2024-09-29 16:47:33.509065] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:11.995 [2024-09-29 16:47:33.509079] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:11.995 [2024-09-29 16:47:33.509091] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:09:11.995 [2024-09-29 16:47:33.509147] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:11.995 pt1 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.995 "name": "raid_bdev1", 00:09:11.995 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:11.995 "strip_size_kb": 0, 00:09:11.995 "state": "configuring", 00:09:11.995 "raid_level": "raid1", 00:09:11.995 "superblock": true, 00:09:11.995 "num_base_bdevs": 3, 00:09:11.995 "num_base_bdevs_discovered": 1, 00:09:11.995 "num_base_bdevs_operational": 2, 00:09:11.995 "base_bdevs_list": [ 00:09:11.995 { 00:09:11.995 "name": null, 00:09:11.995 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:11.995 "is_configured": false, 00:09:11.995 "data_offset": 2048, 00:09:11.995 "data_size": 63488 00:09:11.995 }, 00:09:11.995 { 00:09:11.995 "name": "pt2", 00:09:11.995 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:11.995 "is_configured": true, 00:09:11.995 "data_offset": 2048, 00:09:11.995 "data_size": 63488 00:09:11.995 }, 00:09:11.995 { 00:09:11.995 "name": null, 00:09:11.995 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:11.995 "is_configured": false, 00:09:11.995 "data_offset": 2048, 00:09:11.995 "data_size": 63488 00:09:11.995 } 00:09:11.995 ] 00:09:11.995 }' 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.995 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.563 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:09:12.563 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.563 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.563 16:47:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:12.563 16:47:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.563 [2024-09-29 16:47:34.013830] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:12.563 [2024-09-29 16:47:34.013943] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:12.563 [2024-09-29 16:47:34.013978] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:09:12.563 [2024-09-29 16:47:34.014008] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:12.563 [2024-09-29 16:47:34.014422] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:12.563 [2024-09-29 16:47:34.014484] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:12.563 [2024-09-29 16:47:34.014596] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:12.563 [2024-09-29 16:47:34.014650] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:12.563 [2024-09-29 16:47:34.014787] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:09:12.563 [2024-09-29 16:47:34.014827] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:12.563 [2024-09-29 16:47:34.015117] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:09:12.563 [2024-09-29 16:47:34.015311] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:09:12.563 [2024-09-29 16:47:34.015359] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:09:12.563 [2024-09-29 16:47:34.015543] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:12.563 pt3 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.563 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.563 "name": "raid_bdev1", 00:09:12.563 "uuid": "de86c596-e504-44c3-9c8c-d34e57b67985", 00:09:12.563 "strip_size_kb": 0, 00:09:12.563 "state": "online", 00:09:12.563 "raid_level": "raid1", 00:09:12.563 "superblock": true, 00:09:12.563 "num_base_bdevs": 3, 00:09:12.563 "num_base_bdevs_discovered": 2, 00:09:12.563 "num_base_bdevs_operational": 2, 00:09:12.563 "base_bdevs_list": [ 00:09:12.563 { 00:09:12.563 "name": null, 00:09:12.563 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:12.563 "is_configured": false, 00:09:12.563 "data_offset": 2048, 00:09:12.563 "data_size": 63488 00:09:12.563 }, 00:09:12.563 { 00:09:12.563 "name": "pt2", 00:09:12.563 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:12.563 "is_configured": true, 00:09:12.563 "data_offset": 2048, 00:09:12.563 "data_size": 63488 00:09:12.564 }, 00:09:12.564 { 00:09:12.564 "name": "pt3", 00:09:12.564 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:12.564 "is_configured": true, 00:09:12.564 "data_offset": 2048, 00:09:12.564 "data_size": 63488 00:09:12.564 } 00:09:12.564 ] 00:09:12.564 }' 00:09:12.564 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.564 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.823 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:12.823 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.823 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:12.823 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:12.823 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:13.083 [2024-09-29 16:47:34.505218] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' de86c596-e504-44c3-9c8c-d34e57b67985 '!=' de86c596-e504-44c3-9c8c-d34e57b67985 ']' 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79361 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 79361 ']' 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 79361 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79361 00:09:13.083 killing process with pid 79361 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79361' 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 79361 00:09:13.083 [2024-09-29 16:47:34.589907] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:13.083 [2024-09-29 16:47:34.589981] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:13.083 [2024-09-29 16:47:34.590039] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:13.083 [2024-09-29 16:47:34.590049] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:09:13.083 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 79361 00:09:13.083 [2024-09-29 16:47:34.622702] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:13.343 16:47:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:13.343 00:09:13.343 real 0m6.382s 00:09:13.343 user 0m10.784s 00:09:13.343 sys 0m1.215s 00:09:13.343 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:13.343 16:47:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.343 ************************************ 00:09:13.343 END TEST raid_superblock_test 00:09:13.343 ************************************ 00:09:13.343 16:47:34 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:09:13.343 16:47:34 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:13.343 16:47:34 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:13.343 16:47:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:13.343 ************************************ 00:09:13.343 START TEST raid_read_error_test 00:09:13.343 ************************************ 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 read 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.2VdSkOQ6Yr 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79790 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79790 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 79790 ']' 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:13.343 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:13.343 16:47:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.603 [2024-09-29 16:47:35.030902] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:13.603 [2024-09-29 16:47:35.031125] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79790 ] 00:09:13.603 [2024-09-29 16:47:35.175641] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:13.603 [2024-09-29 16:47:35.222032] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:13.603 [2024-09-29 16:47:35.263793] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:13.603 [2024-09-29 16:47:35.263850] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.173 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:14.173 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:14.173 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:14.173 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:14.173 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.173 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 BaseBdev1_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 true 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 [2024-09-29 16:47:35.873903] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:14.433 [2024-09-29 16:47:35.873964] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.433 [2024-09-29 16:47:35.873992] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:14.433 [2024-09-29 16:47:35.874001] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.433 [2024-09-29 16:47:35.876193] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.433 [2024-09-29 16:47:35.876227] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:14.433 BaseBdev1 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 BaseBdev2_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 true 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 [2024-09-29 16:47:35.920689] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:14.433 [2024-09-29 16:47:35.920745] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.433 [2024-09-29 16:47:35.920766] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:14.433 [2024-09-29 16:47:35.920773] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.433 [2024-09-29 16:47:35.922846] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.433 [2024-09-29 16:47:35.922876] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:14.433 BaseBdev2 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 BaseBdev3_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 true 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.433 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.433 [2024-09-29 16:47:35.949400] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:14.433 [2024-09-29 16:47:35.949445] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.433 [2024-09-29 16:47:35.949464] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:14.434 [2024-09-29 16:47:35.949472] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.434 [2024-09-29 16:47:35.951461] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.434 [2024-09-29 16:47:35.951497] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:14.434 BaseBdev3 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.434 [2024-09-29 16:47:35.957482] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:14.434 [2024-09-29 16:47:35.959243] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:14.434 [2024-09-29 16:47:35.959316] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:14.434 [2024-09-29 16:47:35.959482] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:14.434 [2024-09-29 16:47:35.959497] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:14.434 [2024-09-29 16:47:35.959742] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:14.434 [2024-09-29 16:47:35.959893] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:14.434 [2024-09-29 16:47:35.959953] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:14.434 [2024-09-29 16:47:35.960110] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.434 16:47:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.434 16:47:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.434 "name": "raid_bdev1", 00:09:14.434 "uuid": "31b93cfd-f44b-41e7-9ff8-e0539b902f31", 00:09:14.434 "strip_size_kb": 0, 00:09:14.434 "state": "online", 00:09:14.434 "raid_level": "raid1", 00:09:14.434 "superblock": true, 00:09:14.434 "num_base_bdevs": 3, 00:09:14.434 "num_base_bdevs_discovered": 3, 00:09:14.434 "num_base_bdevs_operational": 3, 00:09:14.434 "base_bdevs_list": [ 00:09:14.434 { 00:09:14.434 "name": "BaseBdev1", 00:09:14.434 "uuid": "69afad42-22c7-5096-b836-c23af2bec20f", 00:09:14.434 "is_configured": true, 00:09:14.434 "data_offset": 2048, 00:09:14.434 "data_size": 63488 00:09:14.434 }, 00:09:14.434 { 00:09:14.434 "name": "BaseBdev2", 00:09:14.434 "uuid": "0428c07e-d759-54f2-ba26-7e7b58fe357f", 00:09:14.434 "is_configured": true, 00:09:14.434 "data_offset": 2048, 00:09:14.434 "data_size": 63488 00:09:14.434 }, 00:09:14.434 { 00:09:14.434 "name": "BaseBdev3", 00:09:14.434 "uuid": "01b46119-fa16-5007-82a2-459de3fc91e4", 00:09:14.434 "is_configured": true, 00:09:14.434 "data_offset": 2048, 00:09:14.434 "data_size": 63488 00:09:14.434 } 00:09:14.434 ] 00:09:14.434 }' 00:09:14.434 16:47:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.434 16:47:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.002 16:47:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:15.002 16:47:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:15.002 [2024-09-29 16:47:36.460946] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.943 "name": "raid_bdev1", 00:09:15.943 "uuid": "31b93cfd-f44b-41e7-9ff8-e0539b902f31", 00:09:15.943 "strip_size_kb": 0, 00:09:15.943 "state": "online", 00:09:15.943 "raid_level": "raid1", 00:09:15.943 "superblock": true, 00:09:15.943 "num_base_bdevs": 3, 00:09:15.943 "num_base_bdevs_discovered": 3, 00:09:15.943 "num_base_bdevs_operational": 3, 00:09:15.943 "base_bdevs_list": [ 00:09:15.943 { 00:09:15.943 "name": "BaseBdev1", 00:09:15.943 "uuid": "69afad42-22c7-5096-b836-c23af2bec20f", 00:09:15.943 "is_configured": true, 00:09:15.943 "data_offset": 2048, 00:09:15.943 "data_size": 63488 00:09:15.943 }, 00:09:15.943 { 00:09:15.943 "name": "BaseBdev2", 00:09:15.943 "uuid": "0428c07e-d759-54f2-ba26-7e7b58fe357f", 00:09:15.943 "is_configured": true, 00:09:15.943 "data_offset": 2048, 00:09:15.943 "data_size": 63488 00:09:15.943 }, 00:09:15.943 { 00:09:15.943 "name": "BaseBdev3", 00:09:15.943 "uuid": "01b46119-fa16-5007-82a2-459de3fc91e4", 00:09:15.943 "is_configured": true, 00:09:15.943 "data_offset": 2048, 00:09:15.943 "data_size": 63488 00:09:15.943 } 00:09:15.943 ] 00:09:15.943 }' 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.943 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.204 [2024-09-29 16:47:37.806563] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:16.204 [2024-09-29 16:47:37.806599] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:16.204 [2024-09-29 16:47:37.809111] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:16.204 [2024-09-29 16:47:37.809162] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:16.204 [2024-09-29 16:47:37.809260] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:16.204 [2024-09-29 16:47:37.809272] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:16.204 { 00:09:16.204 "results": [ 00:09:16.204 { 00:09:16.204 "job": "raid_bdev1", 00:09:16.204 "core_mask": "0x1", 00:09:16.204 "workload": "randrw", 00:09:16.204 "percentage": 50, 00:09:16.204 "status": "finished", 00:09:16.204 "queue_depth": 1, 00:09:16.204 "io_size": 131072, 00:09:16.204 "runtime": 1.346441, 00:09:16.204 "iops": 14869.57096523353, 00:09:16.204 "mibps": 1858.6963706541912, 00:09:16.204 "io_failed": 0, 00:09:16.204 "io_timeout": 0, 00:09:16.204 "avg_latency_us": 64.77442807323052, 00:09:16.204 "min_latency_us": 21.799126637554586, 00:09:16.204 "max_latency_us": 1416.6078602620087 00:09:16.204 } 00:09:16.204 ], 00:09:16.204 "core_count": 1 00:09:16.204 } 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79790 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 79790 ']' 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 79790 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79790 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:16.204 killing process with pid 79790 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79790' 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 79790 00:09:16.204 [2024-09-29 16:47:37.857306] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:16.204 16:47:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 79790 00:09:16.464 [2024-09-29 16:47:37.883033] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.2VdSkOQ6Yr 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:16.464 ************************************ 00:09:16.464 END TEST raid_read_error_test 00:09:16.464 ************************************ 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:16.464 00:09:16.464 real 0m3.198s 00:09:16.464 user 0m3.984s 00:09:16.464 sys 0m0.539s 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:16.464 16:47:38 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.724 16:47:38 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:09:16.724 16:47:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:16.724 16:47:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:16.724 16:47:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:16.724 ************************************ 00:09:16.724 START TEST raid_write_error_test 00:09:16.724 ************************************ 00:09:16.724 16:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 write 00:09:16.724 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.mNykpmdb9F 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79919 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79919 00:09:16.725 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 79919 ']' 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:16.725 16:47:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.725 [2024-09-29 16:47:38.298116] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:16.725 [2024-09-29 16:47:38.298241] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79919 ] 00:09:16.984 [2024-09-29 16:47:38.424111] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:16.984 [2024-09-29 16:47:38.468593] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:16.984 [2024-09-29 16:47:38.510328] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:16.984 [2024-09-29 16:47:38.510366] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.554 BaseBdev1_malloc 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.554 true 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.554 [2024-09-29 16:47:39.152229] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:17.554 [2024-09-29 16:47:39.152282] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:17.554 [2024-09-29 16:47:39.152306] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:17.554 [2024-09-29 16:47:39.152315] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:17.554 [2024-09-29 16:47:39.154420] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:17.554 [2024-09-29 16:47:39.154506] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:17.554 BaseBdev1 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.554 BaseBdev2_malloc 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.554 true 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.554 [2024-09-29 16:47:39.208472] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:17.554 [2024-09-29 16:47:39.208549] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:17.554 [2024-09-29 16:47:39.208579] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:17.554 [2024-09-29 16:47:39.208592] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:17.554 [2024-09-29 16:47:39.211550] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:17.554 [2024-09-29 16:47:39.211658] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:17.554 BaseBdev2 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.554 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.814 BaseBdev3_malloc 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.814 true 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.814 [2024-09-29 16:47:39.249213] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:17.814 [2024-09-29 16:47:39.249310] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:17.814 [2024-09-29 16:47:39.249333] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:17.814 [2024-09-29 16:47:39.249342] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:17.814 [2024-09-29 16:47:39.251346] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:17.814 [2024-09-29 16:47:39.251382] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:17.814 BaseBdev3 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.814 [2024-09-29 16:47:39.261271] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:17.814 [2024-09-29 16:47:39.263049] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:17.814 [2024-09-29 16:47:39.263123] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:17.814 [2024-09-29 16:47:39.263306] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:17.814 [2024-09-29 16:47:39.263320] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:17.814 [2024-09-29 16:47:39.263543] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:17.814 [2024-09-29 16:47:39.263684] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:17.814 [2024-09-29 16:47:39.263693] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:17.814 [2024-09-29 16:47:39.263819] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:17.814 "name": "raid_bdev1", 00:09:17.814 "uuid": "ae8c29ae-ef9d-4b72-97f8-677ed42d2722", 00:09:17.814 "strip_size_kb": 0, 00:09:17.814 "state": "online", 00:09:17.814 "raid_level": "raid1", 00:09:17.814 "superblock": true, 00:09:17.814 "num_base_bdevs": 3, 00:09:17.814 "num_base_bdevs_discovered": 3, 00:09:17.814 "num_base_bdevs_operational": 3, 00:09:17.814 "base_bdevs_list": [ 00:09:17.814 { 00:09:17.814 "name": "BaseBdev1", 00:09:17.814 "uuid": "faeffa97-a4eb-576e-97df-9ef13c055a02", 00:09:17.814 "is_configured": true, 00:09:17.814 "data_offset": 2048, 00:09:17.814 "data_size": 63488 00:09:17.814 }, 00:09:17.814 { 00:09:17.814 "name": "BaseBdev2", 00:09:17.814 "uuid": "5eacd4a1-c13a-55cc-8db9-81788656abbb", 00:09:17.814 "is_configured": true, 00:09:17.814 "data_offset": 2048, 00:09:17.814 "data_size": 63488 00:09:17.814 }, 00:09:17.814 { 00:09:17.814 "name": "BaseBdev3", 00:09:17.814 "uuid": "876cf468-45fc-5234-875c-d91c247a1c2e", 00:09:17.814 "is_configured": true, 00:09:17.814 "data_offset": 2048, 00:09:17.814 "data_size": 63488 00:09:17.814 } 00:09:17.814 ] 00:09:17.814 }' 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:17.814 16:47:39 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.074 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:18.074 16:47:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:18.333 [2024-09-29 16:47:39.772782] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.274 [2024-09-29 16:47:40.715446] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:19.274 [2024-09-29 16:47:40.715590] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:19.274 [2024-09-29 16:47:40.715843] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002600 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:19.274 "name": "raid_bdev1", 00:09:19.274 "uuid": "ae8c29ae-ef9d-4b72-97f8-677ed42d2722", 00:09:19.274 "strip_size_kb": 0, 00:09:19.274 "state": "online", 00:09:19.274 "raid_level": "raid1", 00:09:19.274 "superblock": true, 00:09:19.274 "num_base_bdevs": 3, 00:09:19.274 "num_base_bdevs_discovered": 2, 00:09:19.274 "num_base_bdevs_operational": 2, 00:09:19.274 "base_bdevs_list": [ 00:09:19.274 { 00:09:19.274 "name": null, 00:09:19.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:19.274 "is_configured": false, 00:09:19.274 "data_offset": 0, 00:09:19.274 "data_size": 63488 00:09:19.274 }, 00:09:19.274 { 00:09:19.274 "name": "BaseBdev2", 00:09:19.274 "uuid": "5eacd4a1-c13a-55cc-8db9-81788656abbb", 00:09:19.274 "is_configured": true, 00:09:19.274 "data_offset": 2048, 00:09:19.274 "data_size": 63488 00:09:19.274 }, 00:09:19.274 { 00:09:19.274 "name": "BaseBdev3", 00:09:19.274 "uuid": "876cf468-45fc-5234-875c-d91c247a1c2e", 00:09:19.274 "is_configured": true, 00:09:19.274 "data_offset": 2048, 00:09:19.274 "data_size": 63488 00:09:19.274 } 00:09:19.274 ] 00:09:19.274 }' 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:19.274 16:47:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.534 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:19.534 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:19.534 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.534 [2024-09-29 16:47:41.101451] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:19.534 [2024-09-29 16:47:41.101482] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:19.534 [2024-09-29 16:47:41.103968] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:19.534 [2024-09-29 16:47:41.104013] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:19.534 [2024-09-29 16:47:41.104095] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:19.534 [2024-09-29 16:47:41.104104] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:19.534 { 00:09:19.534 "results": [ 00:09:19.534 { 00:09:19.535 "job": "raid_bdev1", 00:09:19.535 "core_mask": "0x1", 00:09:19.535 "workload": "randrw", 00:09:19.535 "percentage": 50, 00:09:19.535 "status": "finished", 00:09:19.535 "queue_depth": 1, 00:09:19.535 "io_size": 131072, 00:09:19.535 "runtime": 1.32939, 00:09:19.535 "iops": 16374.42736894365, 00:09:19.535 "mibps": 2046.8034211179563, 00:09:19.535 "io_failed": 0, 00:09:19.535 "io_timeout": 0, 00:09:19.535 "avg_latency_us": 58.56730684358595, 00:09:19.535 "min_latency_us": 21.799126637554586, 00:09:19.535 "max_latency_us": 1402.2986899563318 00:09:19.535 } 00:09:19.535 ], 00:09:19.535 "core_count": 1 00:09:19.535 } 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79919 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 79919 ']' 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 79919 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79919 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79919' 00:09:19.535 killing process with pid 79919 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 79919 00:09:19.535 [2024-09-29 16:47:41.151690] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:19.535 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 79919 00:09:19.535 [2024-09-29 16:47:41.176249] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.mNykpmdb9F 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:19.794 00:09:19.794 real 0m3.221s 00:09:19.794 user 0m4.007s 00:09:19.794 sys 0m0.503s 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:19.794 16:47:41 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.794 ************************************ 00:09:19.794 END TEST raid_write_error_test 00:09:19.794 ************************************ 00:09:20.055 16:47:41 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:20.055 16:47:41 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:20.055 16:47:41 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:09:20.055 16:47:41 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:20.055 16:47:41 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:20.055 16:47:41 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:20.055 ************************************ 00:09:20.055 START TEST raid_state_function_test 00:09:20.055 ************************************ 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 false 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:20.055 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80046 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80046' 00:09:20.056 Process raid pid: 80046 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80046 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 80046 ']' 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:20.056 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:20.056 16:47:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.056 [2024-09-29 16:47:41.583590] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:20.056 [2024-09-29 16:47:41.583706] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:20.056 [2024-09-29 16:47:41.726464] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:20.316 [2024-09-29 16:47:41.774033] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:20.316 [2024-09-29 16:47:41.816294] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:20.316 [2024-09-29 16:47:41.816335] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.885 [2024-09-29 16:47:42.409547] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:20.885 [2024-09-29 16:47:42.409603] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:20.885 [2024-09-29 16:47:42.409615] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:20.885 [2024-09-29 16:47:42.409625] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:20.885 [2024-09-29 16:47:42.409631] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:20.885 [2024-09-29 16:47:42.409642] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:20.885 [2024-09-29 16:47:42.409648] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:20.885 [2024-09-29 16:47:42.409656] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.885 "name": "Existed_Raid", 00:09:20.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.885 "strip_size_kb": 64, 00:09:20.885 "state": "configuring", 00:09:20.885 "raid_level": "raid0", 00:09:20.885 "superblock": false, 00:09:20.885 "num_base_bdevs": 4, 00:09:20.885 "num_base_bdevs_discovered": 0, 00:09:20.885 "num_base_bdevs_operational": 4, 00:09:20.885 "base_bdevs_list": [ 00:09:20.885 { 00:09:20.885 "name": "BaseBdev1", 00:09:20.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.885 "is_configured": false, 00:09:20.885 "data_offset": 0, 00:09:20.885 "data_size": 0 00:09:20.885 }, 00:09:20.885 { 00:09:20.885 "name": "BaseBdev2", 00:09:20.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.885 "is_configured": false, 00:09:20.885 "data_offset": 0, 00:09:20.885 "data_size": 0 00:09:20.885 }, 00:09:20.885 { 00:09:20.885 "name": "BaseBdev3", 00:09:20.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.885 "is_configured": false, 00:09:20.885 "data_offset": 0, 00:09:20.885 "data_size": 0 00:09:20.885 }, 00:09:20.885 { 00:09:20.885 "name": "BaseBdev4", 00:09:20.885 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:20.885 "is_configured": false, 00:09:20.885 "data_offset": 0, 00:09:20.885 "data_size": 0 00:09:20.885 } 00:09:20.885 ] 00:09:20.885 }' 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.885 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.455 [2024-09-29 16:47:42.836689] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:21.455 [2024-09-29 16:47:42.836792] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.455 [2024-09-29 16:47:42.848689] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:21.455 [2024-09-29 16:47:42.848777] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:21.455 [2024-09-29 16:47:42.848804] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:21.455 [2024-09-29 16:47:42.848826] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:21.455 [2024-09-29 16:47:42.848843] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:21.455 [2024-09-29 16:47:42.848863] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:21.455 [2024-09-29 16:47:42.848879] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:21.455 [2024-09-29 16:47:42.848914] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.455 [2024-09-29 16:47:42.869380] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:21.455 BaseBdev1 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.455 [ 00:09:21.455 { 00:09:21.455 "name": "BaseBdev1", 00:09:21.455 "aliases": [ 00:09:21.455 "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2" 00:09:21.455 ], 00:09:21.455 "product_name": "Malloc disk", 00:09:21.455 "block_size": 512, 00:09:21.455 "num_blocks": 65536, 00:09:21.455 "uuid": "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2", 00:09:21.455 "assigned_rate_limits": { 00:09:21.455 "rw_ios_per_sec": 0, 00:09:21.455 "rw_mbytes_per_sec": 0, 00:09:21.455 "r_mbytes_per_sec": 0, 00:09:21.455 "w_mbytes_per_sec": 0 00:09:21.455 }, 00:09:21.455 "claimed": true, 00:09:21.455 "claim_type": "exclusive_write", 00:09:21.455 "zoned": false, 00:09:21.455 "supported_io_types": { 00:09:21.455 "read": true, 00:09:21.455 "write": true, 00:09:21.455 "unmap": true, 00:09:21.455 "flush": true, 00:09:21.455 "reset": true, 00:09:21.455 "nvme_admin": false, 00:09:21.455 "nvme_io": false, 00:09:21.455 "nvme_io_md": false, 00:09:21.455 "write_zeroes": true, 00:09:21.455 "zcopy": true, 00:09:21.455 "get_zone_info": false, 00:09:21.455 "zone_management": false, 00:09:21.455 "zone_append": false, 00:09:21.455 "compare": false, 00:09:21.455 "compare_and_write": false, 00:09:21.455 "abort": true, 00:09:21.455 "seek_hole": false, 00:09:21.455 "seek_data": false, 00:09:21.455 "copy": true, 00:09:21.455 "nvme_iov_md": false 00:09:21.455 }, 00:09:21.455 "memory_domains": [ 00:09:21.455 { 00:09:21.455 "dma_device_id": "system", 00:09:21.455 "dma_device_type": 1 00:09:21.455 }, 00:09:21.455 { 00:09:21.455 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:21.455 "dma_device_type": 2 00:09:21.455 } 00:09:21.455 ], 00:09:21.455 "driver_specific": {} 00:09:21.455 } 00:09:21.455 ] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.455 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.455 "name": "Existed_Raid", 00:09:21.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.455 "strip_size_kb": 64, 00:09:21.455 "state": "configuring", 00:09:21.455 "raid_level": "raid0", 00:09:21.455 "superblock": false, 00:09:21.455 "num_base_bdevs": 4, 00:09:21.455 "num_base_bdevs_discovered": 1, 00:09:21.455 "num_base_bdevs_operational": 4, 00:09:21.455 "base_bdevs_list": [ 00:09:21.455 { 00:09:21.455 "name": "BaseBdev1", 00:09:21.455 "uuid": "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2", 00:09:21.455 "is_configured": true, 00:09:21.455 "data_offset": 0, 00:09:21.455 "data_size": 65536 00:09:21.455 }, 00:09:21.455 { 00:09:21.455 "name": "BaseBdev2", 00:09:21.455 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.455 "is_configured": false, 00:09:21.455 "data_offset": 0, 00:09:21.455 "data_size": 0 00:09:21.455 }, 00:09:21.455 { 00:09:21.455 "name": "BaseBdev3", 00:09:21.456 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.456 "is_configured": false, 00:09:21.456 "data_offset": 0, 00:09:21.456 "data_size": 0 00:09:21.456 }, 00:09:21.456 { 00:09:21.456 "name": "BaseBdev4", 00:09:21.456 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.456 "is_configured": false, 00:09:21.456 "data_offset": 0, 00:09:21.456 "data_size": 0 00:09:21.456 } 00:09:21.456 ] 00:09:21.456 }' 00:09:21.456 16:47:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.456 16:47:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.716 [2024-09-29 16:47:43.344605] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:21.716 [2024-09-29 16:47:43.344690] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.716 [2024-09-29 16:47:43.356627] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:21.716 [2024-09-29 16:47:43.358461] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:21.716 [2024-09-29 16:47:43.358549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:21.716 [2024-09-29 16:47:43.358576] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:21.716 [2024-09-29 16:47:43.358597] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:21.716 [2024-09-29 16:47:43.358614] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:21.716 [2024-09-29 16:47:43.358634] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:21.716 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.976 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:21.976 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:21.976 "name": "Existed_Raid", 00:09:21.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.976 "strip_size_kb": 64, 00:09:21.976 "state": "configuring", 00:09:21.976 "raid_level": "raid0", 00:09:21.976 "superblock": false, 00:09:21.976 "num_base_bdevs": 4, 00:09:21.976 "num_base_bdevs_discovered": 1, 00:09:21.976 "num_base_bdevs_operational": 4, 00:09:21.976 "base_bdevs_list": [ 00:09:21.976 { 00:09:21.976 "name": "BaseBdev1", 00:09:21.976 "uuid": "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2", 00:09:21.976 "is_configured": true, 00:09:21.976 "data_offset": 0, 00:09:21.976 "data_size": 65536 00:09:21.976 }, 00:09:21.976 { 00:09:21.976 "name": "BaseBdev2", 00:09:21.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.976 "is_configured": false, 00:09:21.976 "data_offset": 0, 00:09:21.976 "data_size": 0 00:09:21.976 }, 00:09:21.976 { 00:09:21.976 "name": "BaseBdev3", 00:09:21.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.976 "is_configured": false, 00:09:21.976 "data_offset": 0, 00:09:21.976 "data_size": 0 00:09:21.976 }, 00:09:21.976 { 00:09:21.976 "name": "BaseBdev4", 00:09:21.976 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:21.976 "is_configured": false, 00:09:21.976 "data_offset": 0, 00:09:21.976 "data_size": 0 00:09:21.976 } 00:09:21.976 ] 00:09:21.976 }' 00:09:21.976 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:21.976 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.236 [2024-09-29 16:47:43.783307] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:22.236 BaseBdev2 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.236 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.237 [ 00:09:22.237 { 00:09:22.237 "name": "BaseBdev2", 00:09:22.237 "aliases": [ 00:09:22.237 "e8b05d89-c8fe-4106-94d1-d62ab21c57d8" 00:09:22.237 ], 00:09:22.237 "product_name": "Malloc disk", 00:09:22.237 "block_size": 512, 00:09:22.237 "num_blocks": 65536, 00:09:22.237 "uuid": "e8b05d89-c8fe-4106-94d1-d62ab21c57d8", 00:09:22.237 "assigned_rate_limits": { 00:09:22.237 "rw_ios_per_sec": 0, 00:09:22.237 "rw_mbytes_per_sec": 0, 00:09:22.237 "r_mbytes_per_sec": 0, 00:09:22.237 "w_mbytes_per_sec": 0 00:09:22.237 }, 00:09:22.237 "claimed": true, 00:09:22.237 "claim_type": "exclusive_write", 00:09:22.237 "zoned": false, 00:09:22.237 "supported_io_types": { 00:09:22.237 "read": true, 00:09:22.237 "write": true, 00:09:22.237 "unmap": true, 00:09:22.237 "flush": true, 00:09:22.237 "reset": true, 00:09:22.237 "nvme_admin": false, 00:09:22.237 "nvme_io": false, 00:09:22.237 "nvme_io_md": false, 00:09:22.237 "write_zeroes": true, 00:09:22.237 "zcopy": true, 00:09:22.237 "get_zone_info": false, 00:09:22.237 "zone_management": false, 00:09:22.237 "zone_append": false, 00:09:22.237 "compare": false, 00:09:22.237 "compare_and_write": false, 00:09:22.237 "abort": true, 00:09:22.237 "seek_hole": false, 00:09:22.237 "seek_data": false, 00:09:22.237 "copy": true, 00:09:22.237 "nvme_iov_md": false 00:09:22.237 }, 00:09:22.237 "memory_domains": [ 00:09:22.237 { 00:09:22.237 "dma_device_id": "system", 00:09:22.237 "dma_device_type": 1 00:09:22.237 }, 00:09:22.237 { 00:09:22.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.237 "dma_device_type": 2 00:09:22.237 } 00:09:22.237 ], 00:09:22.237 "driver_specific": {} 00:09:22.237 } 00:09:22.237 ] 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.237 "name": "Existed_Raid", 00:09:22.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.237 "strip_size_kb": 64, 00:09:22.237 "state": "configuring", 00:09:22.237 "raid_level": "raid0", 00:09:22.237 "superblock": false, 00:09:22.237 "num_base_bdevs": 4, 00:09:22.237 "num_base_bdevs_discovered": 2, 00:09:22.237 "num_base_bdevs_operational": 4, 00:09:22.237 "base_bdevs_list": [ 00:09:22.237 { 00:09:22.237 "name": "BaseBdev1", 00:09:22.237 "uuid": "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2", 00:09:22.237 "is_configured": true, 00:09:22.237 "data_offset": 0, 00:09:22.237 "data_size": 65536 00:09:22.237 }, 00:09:22.237 { 00:09:22.237 "name": "BaseBdev2", 00:09:22.237 "uuid": "e8b05d89-c8fe-4106-94d1-d62ab21c57d8", 00:09:22.237 "is_configured": true, 00:09:22.237 "data_offset": 0, 00:09:22.237 "data_size": 65536 00:09:22.237 }, 00:09:22.237 { 00:09:22.237 "name": "BaseBdev3", 00:09:22.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.237 "is_configured": false, 00:09:22.237 "data_offset": 0, 00:09:22.237 "data_size": 0 00:09:22.237 }, 00:09:22.237 { 00:09:22.237 "name": "BaseBdev4", 00:09:22.237 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.237 "is_configured": false, 00:09:22.237 "data_offset": 0, 00:09:22.237 "data_size": 0 00:09:22.237 } 00:09:22.237 ] 00:09:22.237 }' 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.237 16:47:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.838 [2024-09-29 16:47:44.221468] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:22.838 BaseBdev3 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.838 [ 00:09:22.838 { 00:09:22.838 "name": "BaseBdev3", 00:09:22.838 "aliases": [ 00:09:22.838 "6ce81ab6-109d-41e0-97e9-66fb79d85f40" 00:09:22.838 ], 00:09:22.838 "product_name": "Malloc disk", 00:09:22.838 "block_size": 512, 00:09:22.838 "num_blocks": 65536, 00:09:22.838 "uuid": "6ce81ab6-109d-41e0-97e9-66fb79d85f40", 00:09:22.838 "assigned_rate_limits": { 00:09:22.838 "rw_ios_per_sec": 0, 00:09:22.838 "rw_mbytes_per_sec": 0, 00:09:22.838 "r_mbytes_per_sec": 0, 00:09:22.838 "w_mbytes_per_sec": 0 00:09:22.838 }, 00:09:22.838 "claimed": true, 00:09:22.838 "claim_type": "exclusive_write", 00:09:22.838 "zoned": false, 00:09:22.838 "supported_io_types": { 00:09:22.838 "read": true, 00:09:22.838 "write": true, 00:09:22.838 "unmap": true, 00:09:22.838 "flush": true, 00:09:22.838 "reset": true, 00:09:22.838 "nvme_admin": false, 00:09:22.838 "nvme_io": false, 00:09:22.838 "nvme_io_md": false, 00:09:22.838 "write_zeroes": true, 00:09:22.838 "zcopy": true, 00:09:22.838 "get_zone_info": false, 00:09:22.838 "zone_management": false, 00:09:22.838 "zone_append": false, 00:09:22.838 "compare": false, 00:09:22.838 "compare_and_write": false, 00:09:22.838 "abort": true, 00:09:22.838 "seek_hole": false, 00:09:22.838 "seek_data": false, 00:09:22.838 "copy": true, 00:09:22.838 "nvme_iov_md": false 00:09:22.838 }, 00:09:22.838 "memory_domains": [ 00:09:22.838 { 00:09:22.838 "dma_device_id": "system", 00:09:22.838 "dma_device_type": 1 00:09:22.838 }, 00:09:22.838 { 00:09:22.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:22.838 "dma_device_type": 2 00:09:22.838 } 00:09:22.838 ], 00:09:22.838 "driver_specific": {} 00:09:22.838 } 00:09:22.838 ] 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.838 "name": "Existed_Raid", 00:09:22.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.838 "strip_size_kb": 64, 00:09:22.838 "state": "configuring", 00:09:22.838 "raid_level": "raid0", 00:09:22.838 "superblock": false, 00:09:22.838 "num_base_bdevs": 4, 00:09:22.838 "num_base_bdevs_discovered": 3, 00:09:22.838 "num_base_bdevs_operational": 4, 00:09:22.838 "base_bdevs_list": [ 00:09:22.838 { 00:09:22.838 "name": "BaseBdev1", 00:09:22.838 "uuid": "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2", 00:09:22.838 "is_configured": true, 00:09:22.838 "data_offset": 0, 00:09:22.838 "data_size": 65536 00:09:22.838 }, 00:09:22.838 { 00:09:22.838 "name": "BaseBdev2", 00:09:22.838 "uuid": "e8b05d89-c8fe-4106-94d1-d62ab21c57d8", 00:09:22.838 "is_configured": true, 00:09:22.838 "data_offset": 0, 00:09:22.838 "data_size": 65536 00:09:22.838 }, 00:09:22.838 { 00:09:22.838 "name": "BaseBdev3", 00:09:22.838 "uuid": "6ce81ab6-109d-41e0-97e9-66fb79d85f40", 00:09:22.838 "is_configured": true, 00:09:22.838 "data_offset": 0, 00:09:22.838 "data_size": 65536 00:09:22.838 }, 00:09:22.838 { 00:09:22.838 "name": "BaseBdev4", 00:09:22.838 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:22.838 "is_configured": false, 00:09:22.838 "data_offset": 0, 00:09:22.838 "data_size": 0 00:09:22.838 } 00:09:22.838 ] 00:09:22.838 }' 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.838 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.098 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:23.098 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.098 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.098 [2024-09-29 16:47:44.747599] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:23.098 [2024-09-29 16:47:44.747710] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:23.098 [2024-09-29 16:47:44.747754] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:23.098 [2024-09-29 16:47:44.748083] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:23.098 [2024-09-29 16:47:44.748217] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:23.098 [2024-09-29 16:47:44.748237] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:23.098 [2024-09-29 16:47:44.748463] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:23.098 BaseBdev4 00:09:23.098 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.098 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:23.098 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:23.098 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.099 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.358 [ 00:09:23.358 { 00:09:23.358 "name": "BaseBdev4", 00:09:23.358 "aliases": [ 00:09:23.358 "a2bef0ab-448b-48a4-a932-bea3e2b4b892" 00:09:23.358 ], 00:09:23.358 "product_name": "Malloc disk", 00:09:23.358 "block_size": 512, 00:09:23.358 "num_blocks": 65536, 00:09:23.358 "uuid": "a2bef0ab-448b-48a4-a932-bea3e2b4b892", 00:09:23.358 "assigned_rate_limits": { 00:09:23.358 "rw_ios_per_sec": 0, 00:09:23.358 "rw_mbytes_per_sec": 0, 00:09:23.358 "r_mbytes_per_sec": 0, 00:09:23.358 "w_mbytes_per_sec": 0 00:09:23.358 }, 00:09:23.358 "claimed": true, 00:09:23.358 "claim_type": "exclusive_write", 00:09:23.358 "zoned": false, 00:09:23.358 "supported_io_types": { 00:09:23.358 "read": true, 00:09:23.358 "write": true, 00:09:23.358 "unmap": true, 00:09:23.358 "flush": true, 00:09:23.358 "reset": true, 00:09:23.358 "nvme_admin": false, 00:09:23.358 "nvme_io": false, 00:09:23.358 "nvme_io_md": false, 00:09:23.358 "write_zeroes": true, 00:09:23.358 "zcopy": true, 00:09:23.358 "get_zone_info": false, 00:09:23.358 "zone_management": false, 00:09:23.358 "zone_append": false, 00:09:23.358 "compare": false, 00:09:23.358 "compare_and_write": false, 00:09:23.358 "abort": true, 00:09:23.358 "seek_hole": false, 00:09:23.358 "seek_data": false, 00:09:23.358 "copy": true, 00:09:23.358 "nvme_iov_md": false 00:09:23.358 }, 00:09:23.358 "memory_domains": [ 00:09:23.358 { 00:09:23.358 "dma_device_id": "system", 00:09:23.358 "dma_device_type": 1 00:09:23.358 }, 00:09:23.358 { 00:09:23.358 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.358 "dma_device_type": 2 00:09:23.358 } 00:09:23.358 ], 00:09:23.358 "driver_specific": {} 00:09:23.358 } 00:09:23.358 ] 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.358 "name": "Existed_Raid", 00:09:23.358 "uuid": "a300f0c9-3a74-44f3-aa7e-afbdbcaf52c5", 00:09:23.358 "strip_size_kb": 64, 00:09:23.358 "state": "online", 00:09:23.358 "raid_level": "raid0", 00:09:23.358 "superblock": false, 00:09:23.358 "num_base_bdevs": 4, 00:09:23.358 "num_base_bdevs_discovered": 4, 00:09:23.358 "num_base_bdevs_operational": 4, 00:09:23.358 "base_bdevs_list": [ 00:09:23.358 { 00:09:23.358 "name": "BaseBdev1", 00:09:23.358 "uuid": "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2", 00:09:23.358 "is_configured": true, 00:09:23.358 "data_offset": 0, 00:09:23.358 "data_size": 65536 00:09:23.358 }, 00:09:23.358 { 00:09:23.358 "name": "BaseBdev2", 00:09:23.358 "uuid": "e8b05d89-c8fe-4106-94d1-d62ab21c57d8", 00:09:23.358 "is_configured": true, 00:09:23.358 "data_offset": 0, 00:09:23.358 "data_size": 65536 00:09:23.358 }, 00:09:23.358 { 00:09:23.358 "name": "BaseBdev3", 00:09:23.358 "uuid": "6ce81ab6-109d-41e0-97e9-66fb79d85f40", 00:09:23.358 "is_configured": true, 00:09:23.358 "data_offset": 0, 00:09:23.358 "data_size": 65536 00:09:23.358 }, 00:09:23.358 { 00:09:23.358 "name": "BaseBdev4", 00:09:23.358 "uuid": "a2bef0ab-448b-48a4-a932-bea3e2b4b892", 00:09:23.358 "is_configured": true, 00:09:23.358 "data_offset": 0, 00:09:23.358 "data_size": 65536 00:09:23.358 } 00:09:23.358 ] 00:09:23.358 }' 00:09:23.358 16:47:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.359 16:47:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.619 [2024-09-29 16:47:45.243138] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:23.619 "name": "Existed_Raid", 00:09:23.619 "aliases": [ 00:09:23.619 "a300f0c9-3a74-44f3-aa7e-afbdbcaf52c5" 00:09:23.619 ], 00:09:23.619 "product_name": "Raid Volume", 00:09:23.619 "block_size": 512, 00:09:23.619 "num_blocks": 262144, 00:09:23.619 "uuid": "a300f0c9-3a74-44f3-aa7e-afbdbcaf52c5", 00:09:23.619 "assigned_rate_limits": { 00:09:23.619 "rw_ios_per_sec": 0, 00:09:23.619 "rw_mbytes_per_sec": 0, 00:09:23.619 "r_mbytes_per_sec": 0, 00:09:23.619 "w_mbytes_per_sec": 0 00:09:23.619 }, 00:09:23.619 "claimed": false, 00:09:23.619 "zoned": false, 00:09:23.619 "supported_io_types": { 00:09:23.619 "read": true, 00:09:23.619 "write": true, 00:09:23.619 "unmap": true, 00:09:23.619 "flush": true, 00:09:23.619 "reset": true, 00:09:23.619 "nvme_admin": false, 00:09:23.619 "nvme_io": false, 00:09:23.619 "nvme_io_md": false, 00:09:23.619 "write_zeroes": true, 00:09:23.619 "zcopy": false, 00:09:23.619 "get_zone_info": false, 00:09:23.619 "zone_management": false, 00:09:23.619 "zone_append": false, 00:09:23.619 "compare": false, 00:09:23.619 "compare_and_write": false, 00:09:23.619 "abort": false, 00:09:23.619 "seek_hole": false, 00:09:23.619 "seek_data": false, 00:09:23.619 "copy": false, 00:09:23.619 "nvme_iov_md": false 00:09:23.619 }, 00:09:23.619 "memory_domains": [ 00:09:23.619 { 00:09:23.619 "dma_device_id": "system", 00:09:23.619 "dma_device_type": 1 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.619 "dma_device_type": 2 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "dma_device_id": "system", 00:09:23.619 "dma_device_type": 1 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.619 "dma_device_type": 2 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "dma_device_id": "system", 00:09:23.619 "dma_device_type": 1 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.619 "dma_device_type": 2 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "dma_device_id": "system", 00:09:23.619 "dma_device_type": 1 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:23.619 "dma_device_type": 2 00:09:23.619 } 00:09:23.619 ], 00:09:23.619 "driver_specific": { 00:09:23.619 "raid": { 00:09:23.619 "uuid": "a300f0c9-3a74-44f3-aa7e-afbdbcaf52c5", 00:09:23.619 "strip_size_kb": 64, 00:09:23.619 "state": "online", 00:09:23.619 "raid_level": "raid0", 00:09:23.619 "superblock": false, 00:09:23.619 "num_base_bdevs": 4, 00:09:23.619 "num_base_bdevs_discovered": 4, 00:09:23.619 "num_base_bdevs_operational": 4, 00:09:23.619 "base_bdevs_list": [ 00:09:23.619 { 00:09:23.619 "name": "BaseBdev1", 00:09:23.619 "uuid": "680218cf-a1f0-4cfe-9c0f-ce82c7a9aea2", 00:09:23.619 "is_configured": true, 00:09:23.619 "data_offset": 0, 00:09:23.619 "data_size": 65536 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "name": "BaseBdev2", 00:09:23.619 "uuid": "e8b05d89-c8fe-4106-94d1-d62ab21c57d8", 00:09:23.619 "is_configured": true, 00:09:23.619 "data_offset": 0, 00:09:23.619 "data_size": 65536 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "name": "BaseBdev3", 00:09:23.619 "uuid": "6ce81ab6-109d-41e0-97e9-66fb79d85f40", 00:09:23.619 "is_configured": true, 00:09:23.619 "data_offset": 0, 00:09:23.619 "data_size": 65536 00:09:23.619 }, 00:09:23.619 { 00:09:23.619 "name": "BaseBdev4", 00:09:23.619 "uuid": "a2bef0ab-448b-48a4-a932-bea3e2b4b892", 00:09:23.619 "is_configured": true, 00:09:23.619 "data_offset": 0, 00:09:23.619 "data_size": 65536 00:09:23.619 } 00:09:23.619 ] 00:09:23.619 } 00:09:23.619 } 00:09:23.619 }' 00:09:23.619 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:23.880 BaseBdev2 00:09:23.880 BaseBdev3 00:09:23.880 BaseBdev4' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.880 [2024-09-29 16:47:45.506428] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:23.880 [2024-09-29 16:47:45.506502] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:23.880 [2024-09-29 16:47:45.506560] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.880 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.139 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.140 "name": "Existed_Raid", 00:09:24.140 "uuid": "a300f0c9-3a74-44f3-aa7e-afbdbcaf52c5", 00:09:24.140 "strip_size_kb": 64, 00:09:24.140 "state": "offline", 00:09:24.140 "raid_level": "raid0", 00:09:24.140 "superblock": false, 00:09:24.140 "num_base_bdevs": 4, 00:09:24.140 "num_base_bdevs_discovered": 3, 00:09:24.140 "num_base_bdevs_operational": 3, 00:09:24.140 "base_bdevs_list": [ 00:09:24.140 { 00:09:24.140 "name": null, 00:09:24.140 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.140 "is_configured": false, 00:09:24.140 "data_offset": 0, 00:09:24.140 "data_size": 65536 00:09:24.140 }, 00:09:24.140 { 00:09:24.140 "name": "BaseBdev2", 00:09:24.140 "uuid": "e8b05d89-c8fe-4106-94d1-d62ab21c57d8", 00:09:24.140 "is_configured": true, 00:09:24.140 "data_offset": 0, 00:09:24.140 "data_size": 65536 00:09:24.140 }, 00:09:24.140 { 00:09:24.140 "name": "BaseBdev3", 00:09:24.140 "uuid": "6ce81ab6-109d-41e0-97e9-66fb79d85f40", 00:09:24.140 "is_configured": true, 00:09:24.140 "data_offset": 0, 00:09:24.140 "data_size": 65536 00:09:24.140 }, 00:09:24.140 { 00:09:24.140 "name": "BaseBdev4", 00:09:24.140 "uuid": "a2bef0ab-448b-48a4-a932-bea3e2b4b892", 00:09:24.140 "is_configured": true, 00:09:24.140 "data_offset": 0, 00:09:24.140 "data_size": 65536 00:09:24.140 } 00:09:24.140 ] 00:09:24.140 }' 00:09:24.140 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.140 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.400 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:24.400 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:24.400 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.400 16:47:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:24.400 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.400 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.400 16:47:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.400 [2024-09-29 16:47:46.024869] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:24.400 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.661 [2024-09-29 16:47:46.095862] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.661 [2024-09-29 16:47:46.150900] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:24.661 [2024-09-29 16:47:46.150992] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.661 BaseBdev2 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.661 [ 00:09:24.661 { 00:09:24.661 "name": "BaseBdev2", 00:09:24.661 "aliases": [ 00:09:24.661 "1ee00cfb-48c0-45e5-a24a-14afbc6433e6" 00:09:24.661 ], 00:09:24.661 "product_name": "Malloc disk", 00:09:24.661 "block_size": 512, 00:09:24.661 "num_blocks": 65536, 00:09:24.661 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:24.661 "assigned_rate_limits": { 00:09:24.661 "rw_ios_per_sec": 0, 00:09:24.661 "rw_mbytes_per_sec": 0, 00:09:24.661 "r_mbytes_per_sec": 0, 00:09:24.661 "w_mbytes_per_sec": 0 00:09:24.661 }, 00:09:24.661 "claimed": false, 00:09:24.661 "zoned": false, 00:09:24.661 "supported_io_types": { 00:09:24.661 "read": true, 00:09:24.661 "write": true, 00:09:24.661 "unmap": true, 00:09:24.661 "flush": true, 00:09:24.661 "reset": true, 00:09:24.661 "nvme_admin": false, 00:09:24.661 "nvme_io": false, 00:09:24.661 "nvme_io_md": false, 00:09:24.661 "write_zeroes": true, 00:09:24.661 "zcopy": true, 00:09:24.661 "get_zone_info": false, 00:09:24.661 "zone_management": false, 00:09:24.661 "zone_append": false, 00:09:24.661 "compare": false, 00:09:24.661 "compare_and_write": false, 00:09:24.661 "abort": true, 00:09:24.661 "seek_hole": false, 00:09:24.661 "seek_data": false, 00:09:24.661 "copy": true, 00:09:24.661 "nvme_iov_md": false 00:09:24.661 }, 00:09:24.661 "memory_domains": [ 00:09:24.661 { 00:09:24.661 "dma_device_id": "system", 00:09:24.661 "dma_device_type": 1 00:09:24.661 }, 00:09:24.661 { 00:09:24.661 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.661 "dma_device_type": 2 00:09:24.661 } 00:09:24.661 ], 00:09:24.661 "driver_specific": {} 00:09:24.661 } 00:09:24.661 ] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:24.661 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.662 BaseBdev3 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.662 [ 00:09:24.662 { 00:09:24.662 "name": "BaseBdev3", 00:09:24.662 "aliases": [ 00:09:24.662 "96e8ec2c-bb91-4827-864a-180814442520" 00:09:24.662 ], 00:09:24.662 "product_name": "Malloc disk", 00:09:24.662 "block_size": 512, 00:09:24.662 "num_blocks": 65536, 00:09:24.662 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:24.662 "assigned_rate_limits": { 00:09:24.662 "rw_ios_per_sec": 0, 00:09:24.662 "rw_mbytes_per_sec": 0, 00:09:24.662 "r_mbytes_per_sec": 0, 00:09:24.662 "w_mbytes_per_sec": 0 00:09:24.662 }, 00:09:24.662 "claimed": false, 00:09:24.662 "zoned": false, 00:09:24.662 "supported_io_types": { 00:09:24.662 "read": true, 00:09:24.662 "write": true, 00:09:24.662 "unmap": true, 00:09:24.662 "flush": true, 00:09:24.662 "reset": true, 00:09:24.662 "nvme_admin": false, 00:09:24.662 "nvme_io": false, 00:09:24.662 "nvme_io_md": false, 00:09:24.662 "write_zeroes": true, 00:09:24.662 "zcopy": true, 00:09:24.662 "get_zone_info": false, 00:09:24.662 "zone_management": false, 00:09:24.662 "zone_append": false, 00:09:24.662 "compare": false, 00:09:24.662 "compare_and_write": false, 00:09:24.662 "abort": true, 00:09:24.662 "seek_hole": false, 00:09:24.662 "seek_data": false, 00:09:24.662 "copy": true, 00:09:24.662 "nvme_iov_md": false 00:09:24.662 }, 00:09:24.662 "memory_domains": [ 00:09:24.662 { 00:09:24.662 "dma_device_id": "system", 00:09:24.662 "dma_device_type": 1 00:09:24.662 }, 00:09:24.662 { 00:09:24.662 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.662 "dma_device_type": 2 00:09:24.662 } 00:09:24.662 ], 00:09:24.662 "driver_specific": {} 00:09:24.662 } 00:09:24.662 ] 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.662 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.922 BaseBdev4 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.922 [ 00:09:24.922 { 00:09:24.922 "name": "BaseBdev4", 00:09:24.922 "aliases": [ 00:09:24.922 "697fec0c-d8a1-48b1-bee4-667271b589dd" 00:09:24.922 ], 00:09:24.922 "product_name": "Malloc disk", 00:09:24.922 "block_size": 512, 00:09:24.922 "num_blocks": 65536, 00:09:24.922 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:24.922 "assigned_rate_limits": { 00:09:24.922 "rw_ios_per_sec": 0, 00:09:24.922 "rw_mbytes_per_sec": 0, 00:09:24.922 "r_mbytes_per_sec": 0, 00:09:24.922 "w_mbytes_per_sec": 0 00:09:24.922 }, 00:09:24.922 "claimed": false, 00:09:24.922 "zoned": false, 00:09:24.922 "supported_io_types": { 00:09:24.922 "read": true, 00:09:24.922 "write": true, 00:09:24.922 "unmap": true, 00:09:24.922 "flush": true, 00:09:24.922 "reset": true, 00:09:24.922 "nvme_admin": false, 00:09:24.922 "nvme_io": false, 00:09:24.922 "nvme_io_md": false, 00:09:24.922 "write_zeroes": true, 00:09:24.922 "zcopy": true, 00:09:24.922 "get_zone_info": false, 00:09:24.922 "zone_management": false, 00:09:24.922 "zone_append": false, 00:09:24.922 "compare": false, 00:09:24.922 "compare_and_write": false, 00:09:24.922 "abort": true, 00:09:24.922 "seek_hole": false, 00:09:24.922 "seek_data": false, 00:09:24.922 "copy": true, 00:09:24.922 "nvme_iov_md": false 00:09:24.922 }, 00:09:24.922 "memory_domains": [ 00:09:24.922 { 00:09:24.922 "dma_device_id": "system", 00:09:24.922 "dma_device_type": 1 00:09:24.922 }, 00:09:24.922 { 00:09:24.922 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:24.922 "dma_device_type": 2 00:09:24.922 } 00:09:24.922 ], 00:09:24.922 "driver_specific": {} 00:09:24.922 } 00:09:24.922 ] 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:24.922 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.923 [2024-09-29 16:47:46.378015] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:24.923 [2024-09-29 16:47:46.378111] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:24.923 [2024-09-29 16:47:46.378151] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:24.923 [2024-09-29 16:47:46.379967] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:24.923 [2024-09-29 16:47:46.380051] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:24.923 "name": "Existed_Raid", 00:09:24.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.923 "strip_size_kb": 64, 00:09:24.923 "state": "configuring", 00:09:24.923 "raid_level": "raid0", 00:09:24.923 "superblock": false, 00:09:24.923 "num_base_bdevs": 4, 00:09:24.923 "num_base_bdevs_discovered": 3, 00:09:24.923 "num_base_bdevs_operational": 4, 00:09:24.923 "base_bdevs_list": [ 00:09:24.923 { 00:09:24.923 "name": "BaseBdev1", 00:09:24.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:24.923 "is_configured": false, 00:09:24.923 "data_offset": 0, 00:09:24.923 "data_size": 0 00:09:24.923 }, 00:09:24.923 { 00:09:24.923 "name": "BaseBdev2", 00:09:24.923 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:24.923 "is_configured": true, 00:09:24.923 "data_offset": 0, 00:09:24.923 "data_size": 65536 00:09:24.923 }, 00:09:24.923 { 00:09:24.923 "name": "BaseBdev3", 00:09:24.923 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:24.923 "is_configured": true, 00:09:24.923 "data_offset": 0, 00:09:24.923 "data_size": 65536 00:09:24.923 }, 00:09:24.923 { 00:09:24.923 "name": "BaseBdev4", 00:09:24.923 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:24.923 "is_configured": true, 00:09:24.923 "data_offset": 0, 00:09:24.923 "data_size": 65536 00:09:24.923 } 00:09:24.923 ] 00:09:24.923 }' 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:24.923 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.183 [2024-09-29 16:47:46.805291] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.183 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.442 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.442 "name": "Existed_Raid", 00:09:25.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.443 "strip_size_kb": 64, 00:09:25.443 "state": "configuring", 00:09:25.443 "raid_level": "raid0", 00:09:25.443 "superblock": false, 00:09:25.443 "num_base_bdevs": 4, 00:09:25.443 "num_base_bdevs_discovered": 2, 00:09:25.443 "num_base_bdevs_operational": 4, 00:09:25.443 "base_bdevs_list": [ 00:09:25.443 { 00:09:25.443 "name": "BaseBdev1", 00:09:25.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.443 "is_configured": false, 00:09:25.443 "data_offset": 0, 00:09:25.443 "data_size": 0 00:09:25.443 }, 00:09:25.443 { 00:09:25.443 "name": null, 00:09:25.443 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:25.443 "is_configured": false, 00:09:25.443 "data_offset": 0, 00:09:25.443 "data_size": 65536 00:09:25.443 }, 00:09:25.443 { 00:09:25.443 "name": "BaseBdev3", 00:09:25.443 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:25.443 "is_configured": true, 00:09:25.443 "data_offset": 0, 00:09:25.443 "data_size": 65536 00:09:25.443 }, 00:09:25.443 { 00:09:25.443 "name": "BaseBdev4", 00:09:25.443 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:25.443 "is_configured": true, 00:09:25.443 "data_offset": 0, 00:09:25.443 "data_size": 65536 00:09:25.443 } 00:09:25.443 ] 00:09:25.443 }' 00:09:25.443 16:47:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.443 16:47:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.703 [2024-09-29 16:47:47.259410] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:25.703 BaseBdev1 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.703 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.703 [ 00:09:25.703 { 00:09:25.703 "name": "BaseBdev1", 00:09:25.703 "aliases": [ 00:09:25.703 "d4992a07-af83-4662-8615-10eafe1b9c03" 00:09:25.703 ], 00:09:25.703 "product_name": "Malloc disk", 00:09:25.703 "block_size": 512, 00:09:25.703 "num_blocks": 65536, 00:09:25.703 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:25.703 "assigned_rate_limits": { 00:09:25.703 "rw_ios_per_sec": 0, 00:09:25.703 "rw_mbytes_per_sec": 0, 00:09:25.703 "r_mbytes_per_sec": 0, 00:09:25.703 "w_mbytes_per_sec": 0 00:09:25.703 }, 00:09:25.703 "claimed": true, 00:09:25.703 "claim_type": "exclusive_write", 00:09:25.703 "zoned": false, 00:09:25.703 "supported_io_types": { 00:09:25.703 "read": true, 00:09:25.703 "write": true, 00:09:25.703 "unmap": true, 00:09:25.703 "flush": true, 00:09:25.703 "reset": true, 00:09:25.703 "nvme_admin": false, 00:09:25.703 "nvme_io": false, 00:09:25.703 "nvme_io_md": false, 00:09:25.703 "write_zeroes": true, 00:09:25.703 "zcopy": true, 00:09:25.703 "get_zone_info": false, 00:09:25.703 "zone_management": false, 00:09:25.703 "zone_append": false, 00:09:25.703 "compare": false, 00:09:25.703 "compare_and_write": false, 00:09:25.703 "abort": true, 00:09:25.703 "seek_hole": false, 00:09:25.703 "seek_data": false, 00:09:25.703 "copy": true, 00:09:25.703 "nvme_iov_md": false 00:09:25.703 }, 00:09:25.703 "memory_domains": [ 00:09:25.703 { 00:09:25.703 "dma_device_id": "system", 00:09:25.703 "dma_device_type": 1 00:09:25.704 }, 00:09:25.704 { 00:09:25.704 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:25.704 "dma_device_type": 2 00:09:25.704 } 00:09:25.704 ], 00:09:25.704 "driver_specific": {} 00:09:25.704 } 00:09:25.704 ] 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.704 "name": "Existed_Raid", 00:09:25.704 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.704 "strip_size_kb": 64, 00:09:25.704 "state": "configuring", 00:09:25.704 "raid_level": "raid0", 00:09:25.704 "superblock": false, 00:09:25.704 "num_base_bdevs": 4, 00:09:25.704 "num_base_bdevs_discovered": 3, 00:09:25.704 "num_base_bdevs_operational": 4, 00:09:25.704 "base_bdevs_list": [ 00:09:25.704 { 00:09:25.704 "name": "BaseBdev1", 00:09:25.704 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:25.704 "is_configured": true, 00:09:25.704 "data_offset": 0, 00:09:25.704 "data_size": 65536 00:09:25.704 }, 00:09:25.704 { 00:09:25.704 "name": null, 00:09:25.704 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:25.704 "is_configured": false, 00:09:25.704 "data_offset": 0, 00:09:25.704 "data_size": 65536 00:09:25.704 }, 00:09:25.704 { 00:09:25.704 "name": "BaseBdev3", 00:09:25.704 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:25.704 "is_configured": true, 00:09:25.704 "data_offset": 0, 00:09:25.704 "data_size": 65536 00:09:25.704 }, 00:09:25.704 { 00:09:25.704 "name": "BaseBdev4", 00:09:25.704 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:25.704 "is_configured": true, 00:09:25.704 "data_offset": 0, 00:09:25.704 "data_size": 65536 00:09:25.704 } 00:09:25.704 ] 00:09:25.704 }' 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.704 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.274 [2024-09-29 16:47:47.794544] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.274 "name": "Existed_Raid", 00:09:26.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.274 "strip_size_kb": 64, 00:09:26.274 "state": "configuring", 00:09:26.274 "raid_level": "raid0", 00:09:26.274 "superblock": false, 00:09:26.274 "num_base_bdevs": 4, 00:09:26.274 "num_base_bdevs_discovered": 2, 00:09:26.274 "num_base_bdevs_operational": 4, 00:09:26.274 "base_bdevs_list": [ 00:09:26.274 { 00:09:26.274 "name": "BaseBdev1", 00:09:26.274 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:26.274 "is_configured": true, 00:09:26.274 "data_offset": 0, 00:09:26.274 "data_size": 65536 00:09:26.274 }, 00:09:26.274 { 00:09:26.274 "name": null, 00:09:26.274 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:26.274 "is_configured": false, 00:09:26.274 "data_offset": 0, 00:09:26.274 "data_size": 65536 00:09:26.274 }, 00:09:26.274 { 00:09:26.274 "name": null, 00:09:26.274 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:26.274 "is_configured": false, 00:09:26.274 "data_offset": 0, 00:09:26.274 "data_size": 65536 00:09:26.274 }, 00:09:26.274 { 00:09:26.274 "name": "BaseBdev4", 00:09:26.274 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:26.274 "is_configured": true, 00:09:26.274 "data_offset": 0, 00:09:26.274 "data_size": 65536 00:09:26.274 } 00:09:26.274 ] 00:09:26.274 }' 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.274 16:47:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.844 [2024-09-29 16:47:48.249852] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.844 "name": "Existed_Raid", 00:09:26.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.844 "strip_size_kb": 64, 00:09:26.844 "state": "configuring", 00:09:26.844 "raid_level": "raid0", 00:09:26.844 "superblock": false, 00:09:26.844 "num_base_bdevs": 4, 00:09:26.844 "num_base_bdevs_discovered": 3, 00:09:26.844 "num_base_bdevs_operational": 4, 00:09:26.844 "base_bdevs_list": [ 00:09:26.844 { 00:09:26.844 "name": "BaseBdev1", 00:09:26.844 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:26.844 "is_configured": true, 00:09:26.844 "data_offset": 0, 00:09:26.844 "data_size": 65536 00:09:26.844 }, 00:09:26.844 { 00:09:26.844 "name": null, 00:09:26.844 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:26.844 "is_configured": false, 00:09:26.844 "data_offset": 0, 00:09:26.844 "data_size": 65536 00:09:26.844 }, 00:09:26.844 { 00:09:26.844 "name": "BaseBdev3", 00:09:26.844 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:26.844 "is_configured": true, 00:09:26.844 "data_offset": 0, 00:09:26.844 "data_size": 65536 00:09:26.844 }, 00:09:26.844 { 00:09:26.844 "name": "BaseBdev4", 00:09:26.844 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:26.844 "is_configured": true, 00:09:26.844 "data_offset": 0, 00:09:26.844 "data_size": 65536 00:09:26.844 } 00:09:26.844 ] 00:09:26.844 }' 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.844 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.103 [2024-09-29 16:47:48.736991] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.103 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.362 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.362 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.362 "name": "Existed_Raid", 00:09:27.362 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.362 "strip_size_kb": 64, 00:09:27.362 "state": "configuring", 00:09:27.362 "raid_level": "raid0", 00:09:27.362 "superblock": false, 00:09:27.362 "num_base_bdevs": 4, 00:09:27.362 "num_base_bdevs_discovered": 2, 00:09:27.362 "num_base_bdevs_operational": 4, 00:09:27.362 "base_bdevs_list": [ 00:09:27.362 { 00:09:27.362 "name": null, 00:09:27.362 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:27.362 "is_configured": false, 00:09:27.362 "data_offset": 0, 00:09:27.362 "data_size": 65536 00:09:27.362 }, 00:09:27.362 { 00:09:27.362 "name": null, 00:09:27.362 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:27.362 "is_configured": false, 00:09:27.362 "data_offset": 0, 00:09:27.362 "data_size": 65536 00:09:27.362 }, 00:09:27.362 { 00:09:27.362 "name": "BaseBdev3", 00:09:27.362 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:27.362 "is_configured": true, 00:09:27.362 "data_offset": 0, 00:09:27.362 "data_size": 65536 00:09:27.362 }, 00:09:27.362 { 00:09:27.362 "name": "BaseBdev4", 00:09:27.362 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:27.362 "is_configured": true, 00:09:27.362 "data_offset": 0, 00:09:27.362 "data_size": 65536 00:09:27.362 } 00:09:27.362 ] 00:09:27.362 }' 00:09:27.362 16:47:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.362 16:47:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.621 [2024-09-29 16:47:49.254612] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.621 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.622 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.880 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.880 "name": "Existed_Raid", 00:09:27.880 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.880 "strip_size_kb": 64, 00:09:27.880 "state": "configuring", 00:09:27.880 "raid_level": "raid0", 00:09:27.880 "superblock": false, 00:09:27.880 "num_base_bdevs": 4, 00:09:27.880 "num_base_bdevs_discovered": 3, 00:09:27.880 "num_base_bdevs_operational": 4, 00:09:27.880 "base_bdevs_list": [ 00:09:27.880 { 00:09:27.880 "name": null, 00:09:27.880 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:27.880 "is_configured": false, 00:09:27.880 "data_offset": 0, 00:09:27.880 "data_size": 65536 00:09:27.880 }, 00:09:27.880 { 00:09:27.880 "name": "BaseBdev2", 00:09:27.880 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:27.880 "is_configured": true, 00:09:27.880 "data_offset": 0, 00:09:27.880 "data_size": 65536 00:09:27.880 }, 00:09:27.880 { 00:09:27.880 "name": "BaseBdev3", 00:09:27.880 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:27.880 "is_configured": true, 00:09:27.880 "data_offset": 0, 00:09:27.880 "data_size": 65536 00:09:27.880 }, 00:09:27.880 { 00:09:27.880 "name": "BaseBdev4", 00:09:27.880 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:27.880 "is_configured": true, 00:09:27.880 "data_offset": 0, 00:09:27.880 "data_size": 65536 00:09:27.880 } 00:09:27.880 ] 00:09:27.880 }' 00:09:27.880 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.880 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d4992a07-af83-4662-8615-10eafe1b9c03 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.139 [2024-09-29 16:47:49.744660] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:28.139 [2024-09-29 16:47:49.744789] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:28.139 [2024-09-29 16:47:49.744802] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:28.139 [2024-09-29 16:47:49.745080] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:28.139 [2024-09-29 16:47:49.745193] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:28.139 [2024-09-29 16:47:49.745204] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:28.139 [2024-09-29 16:47:49.745372] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:28.139 NewBaseBdev 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:28.139 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.140 [ 00:09:28.140 { 00:09:28.140 "name": "NewBaseBdev", 00:09:28.140 "aliases": [ 00:09:28.140 "d4992a07-af83-4662-8615-10eafe1b9c03" 00:09:28.140 ], 00:09:28.140 "product_name": "Malloc disk", 00:09:28.140 "block_size": 512, 00:09:28.140 "num_blocks": 65536, 00:09:28.140 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:28.140 "assigned_rate_limits": { 00:09:28.140 "rw_ios_per_sec": 0, 00:09:28.140 "rw_mbytes_per_sec": 0, 00:09:28.140 "r_mbytes_per_sec": 0, 00:09:28.140 "w_mbytes_per_sec": 0 00:09:28.140 }, 00:09:28.140 "claimed": true, 00:09:28.140 "claim_type": "exclusive_write", 00:09:28.140 "zoned": false, 00:09:28.140 "supported_io_types": { 00:09:28.140 "read": true, 00:09:28.140 "write": true, 00:09:28.140 "unmap": true, 00:09:28.140 "flush": true, 00:09:28.140 "reset": true, 00:09:28.140 "nvme_admin": false, 00:09:28.140 "nvme_io": false, 00:09:28.140 "nvme_io_md": false, 00:09:28.140 "write_zeroes": true, 00:09:28.140 "zcopy": true, 00:09:28.140 "get_zone_info": false, 00:09:28.140 "zone_management": false, 00:09:28.140 "zone_append": false, 00:09:28.140 "compare": false, 00:09:28.140 "compare_and_write": false, 00:09:28.140 "abort": true, 00:09:28.140 "seek_hole": false, 00:09:28.140 "seek_data": false, 00:09:28.140 "copy": true, 00:09:28.140 "nvme_iov_md": false 00:09:28.140 }, 00:09:28.140 "memory_domains": [ 00:09:28.140 { 00:09:28.140 "dma_device_id": "system", 00:09:28.140 "dma_device_type": 1 00:09:28.140 }, 00:09:28.140 { 00:09:28.140 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.140 "dma_device_type": 2 00:09:28.140 } 00:09:28.140 ], 00:09:28.140 "driver_specific": {} 00:09:28.140 } 00:09:28.140 ] 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.140 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.399 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.399 "name": "Existed_Raid", 00:09:28.400 "uuid": "9f7a670f-3e47-4606-b3ed-6b21c49919c1", 00:09:28.400 "strip_size_kb": 64, 00:09:28.400 "state": "online", 00:09:28.400 "raid_level": "raid0", 00:09:28.400 "superblock": false, 00:09:28.400 "num_base_bdevs": 4, 00:09:28.400 "num_base_bdevs_discovered": 4, 00:09:28.400 "num_base_bdevs_operational": 4, 00:09:28.400 "base_bdevs_list": [ 00:09:28.400 { 00:09:28.400 "name": "NewBaseBdev", 00:09:28.400 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:28.400 "is_configured": true, 00:09:28.400 "data_offset": 0, 00:09:28.400 "data_size": 65536 00:09:28.400 }, 00:09:28.400 { 00:09:28.400 "name": "BaseBdev2", 00:09:28.400 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:28.400 "is_configured": true, 00:09:28.400 "data_offset": 0, 00:09:28.400 "data_size": 65536 00:09:28.400 }, 00:09:28.400 { 00:09:28.400 "name": "BaseBdev3", 00:09:28.400 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:28.400 "is_configured": true, 00:09:28.400 "data_offset": 0, 00:09:28.400 "data_size": 65536 00:09:28.400 }, 00:09:28.400 { 00:09:28.400 "name": "BaseBdev4", 00:09:28.400 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:28.400 "is_configured": true, 00:09:28.400 "data_offset": 0, 00:09:28.400 "data_size": 65536 00:09:28.400 } 00:09:28.400 ] 00:09:28.400 }' 00:09:28.400 16:47:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.400 16:47:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:28.660 [2024-09-29 16:47:50.236178] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:28.660 "name": "Existed_Raid", 00:09:28.660 "aliases": [ 00:09:28.660 "9f7a670f-3e47-4606-b3ed-6b21c49919c1" 00:09:28.660 ], 00:09:28.660 "product_name": "Raid Volume", 00:09:28.660 "block_size": 512, 00:09:28.660 "num_blocks": 262144, 00:09:28.660 "uuid": "9f7a670f-3e47-4606-b3ed-6b21c49919c1", 00:09:28.660 "assigned_rate_limits": { 00:09:28.660 "rw_ios_per_sec": 0, 00:09:28.660 "rw_mbytes_per_sec": 0, 00:09:28.660 "r_mbytes_per_sec": 0, 00:09:28.660 "w_mbytes_per_sec": 0 00:09:28.660 }, 00:09:28.660 "claimed": false, 00:09:28.660 "zoned": false, 00:09:28.660 "supported_io_types": { 00:09:28.660 "read": true, 00:09:28.660 "write": true, 00:09:28.660 "unmap": true, 00:09:28.660 "flush": true, 00:09:28.660 "reset": true, 00:09:28.660 "nvme_admin": false, 00:09:28.660 "nvme_io": false, 00:09:28.660 "nvme_io_md": false, 00:09:28.660 "write_zeroes": true, 00:09:28.660 "zcopy": false, 00:09:28.660 "get_zone_info": false, 00:09:28.660 "zone_management": false, 00:09:28.660 "zone_append": false, 00:09:28.660 "compare": false, 00:09:28.660 "compare_and_write": false, 00:09:28.660 "abort": false, 00:09:28.660 "seek_hole": false, 00:09:28.660 "seek_data": false, 00:09:28.660 "copy": false, 00:09:28.660 "nvme_iov_md": false 00:09:28.660 }, 00:09:28.660 "memory_domains": [ 00:09:28.660 { 00:09:28.660 "dma_device_id": "system", 00:09:28.660 "dma_device_type": 1 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.660 "dma_device_type": 2 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "dma_device_id": "system", 00:09:28.660 "dma_device_type": 1 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.660 "dma_device_type": 2 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "dma_device_id": "system", 00:09:28.660 "dma_device_type": 1 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.660 "dma_device_type": 2 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "dma_device_id": "system", 00:09:28.660 "dma_device_type": 1 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.660 "dma_device_type": 2 00:09:28.660 } 00:09:28.660 ], 00:09:28.660 "driver_specific": { 00:09:28.660 "raid": { 00:09:28.660 "uuid": "9f7a670f-3e47-4606-b3ed-6b21c49919c1", 00:09:28.660 "strip_size_kb": 64, 00:09:28.660 "state": "online", 00:09:28.660 "raid_level": "raid0", 00:09:28.660 "superblock": false, 00:09:28.660 "num_base_bdevs": 4, 00:09:28.660 "num_base_bdevs_discovered": 4, 00:09:28.660 "num_base_bdevs_operational": 4, 00:09:28.660 "base_bdevs_list": [ 00:09:28.660 { 00:09:28.660 "name": "NewBaseBdev", 00:09:28.660 "uuid": "d4992a07-af83-4662-8615-10eafe1b9c03", 00:09:28.660 "is_configured": true, 00:09:28.660 "data_offset": 0, 00:09:28.660 "data_size": 65536 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "name": "BaseBdev2", 00:09:28.660 "uuid": "1ee00cfb-48c0-45e5-a24a-14afbc6433e6", 00:09:28.660 "is_configured": true, 00:09:28.660 "data_offset": 0, 00:09:28.660 "data_size": 65536 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "name": "BaseBdev3", 00:09:28.660 "uuid": "96e8ec2c-bb91-4827-864a-180814442520", 00:09:28.660 "is_configured": true, 00:09:28.660 "data_offset": 0, 00:09:28.660 "data_size": 65536 00:09:28.660 }, 00:09:28.660 { 00:09:28.660 "name": "BaseBdev4", 00:09:28.660 "uuid": "697fec0c-d8a1-48b1-bee4-667271b589dd", 00:09:28.660 "is_configured": true, 00:09:28.660 "data_offset": 0, 00:09:28.660 "data_size": 65536 00:09:28.660 } 00:09:28.660 ] 00:09:28.660 } 00:09:28.660 } 00:09:28.660 }' 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:28.660 BaseBdev2 00:09:28.660 BaseBdev3 00:09:28.660 BaseBdev4' 00:09:28.660 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.920 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.921 [2024-09-29 16:47:50.507443] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:28.921 [2024-09-29 16:47:50.507512] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:28.921 [2024-09-29 16:47:50.507598] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:28.921 [2024-09-29 16:47:50.507675] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:28.921 [2024-09-29 16:47:50.507757] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80046 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 80046 ']' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 80046 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80046 00:09:28.921 killing process with pid 80046 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80046' 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 80046 00:09:28.921 [2024-09-29 16:47:50.546498] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:28.921 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 80046 00:09:28.921 [2024-09-29 16:47:50.585628] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:29.182 16:47:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:29.182 00:09:29.182 real 0m9.324s 00:09:29.182 user 0m15.995s 00:09:29.182 sys 0m1.849s 00:09:29.182 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:29.182 16:47:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.182 ************************************ 00:09:29.182 END TEST raid_state_function_test 00:09:29.182 ************************************ 00:09:29.444 16:47:50 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:09:29.444 16:47:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:29.444 16:47:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:29.444 16:47:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:29.444 ************************************ 00:09:29.444 START TEST raid_state_function_test_sb 00:09:29.444 ************************************ 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 true 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=80690 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80690' 00:09:29.444 Process raid pid: 80690 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 80690 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 80690 ']' 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:29.444 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:29.444 16:47:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:29.444 [2024-09-29 16:47:50.979706] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:29.444 [2024-09-29 16:47:50.979894] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:29.444 [2024-09-29 16:47:51.105151] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:29.704 [2024-09-29 16:47:51.151415] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:29.704 [2024-09-29 16:47:51.193443] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:29.704 [2024-09-29 16:47:51.193551] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:30.273 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:30.273 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:09:30.273 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:30.273 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.273 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.273 [2024-09-29 16:47:51.818924] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:30.273 [2024-09-29 16:47:51.819030] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:30.273 [2024-09-29 16:47:51.819073] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:30.273 [2024-09-29 16:47:51.819098] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:30.273 [2024-09-29 16:47:51.819116] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:30.273 [2024-09-29 16:47:51.819138] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:30.273 [2024-09-29 16:47:51.819155] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:30.273 [2024-09-29 16:47:51.819175] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.274 "name": "Existed_Raid", 00:09:30.274 "uuid": "b3bdfc48-50de-44d3-adb1-126bc35b53b4", 00:09:30.274 "strip_size_kb": 64, 00:09:30.274 "state": "configuring", 00:09:30.274 "raid_level": "raid0", 00:09:30.274 "superblock": true, 00:09:30.274 "num_base_bdevs": 4, 00:09:30.274 "num_base_bdevs_discovered": 0, 00:09:30.274 "num_base_bdevs_operational": 4, 00:09:30.274 "base_bdevs_list": [ 00:09:30.274 { 00:09:30.274 "name": "BaseBdev1", 00:09:30.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.274 "is_configured": false, 00:09:30.274 "data_offset": 0, 00:09:30.274 "data_size": 0 00:09:30.274 }, 00:09:30.274 { 00:09:30.274 "name": "BaseBdev2", 00:09:30.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.274 "is_configured": false, 00:09:30.274 "data_offset": 0, 00:09:30.274 "data_size": 0 00:09:30.274 }, 00:09:30.274 { 00:09:30.274 "name": "BaseBdev3", 00:09:30.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.274 "is_configured": false, 00:09:30.274 "data_offset": 0, 00:09:30.274 "data_size": 0 00:09:30.274 }, 00:09:30.274 { 00:09:30.274 "name": "BaseBdev4", 00:09:30.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.274 "is_configured": false, 00:09:30.274 "data_offset": 0, 00:09:30.274 "data_size": 0 00:09:30.274 } 00:09:30.274 ] 00:09:30.274 }' 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.274 16:47:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.843 [2024-09-29 16:47:52.285963] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:30.843 [2024-09-29 16:47:52.286045] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.843 [2024-09-29 16:47:52.297972] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:30.843 [2024-09-29 16:47:52.298013] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:30.843 [2024-09-29 16:47:52.298021] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:30.843 [2024-09-29 16:47:52.298030] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:30.843 [2024-09-29 16:47:52.298036] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:30.843 [2024-09-29 16:47:52.298044] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:30.843 [2024-09-29 16:47:52.298050] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:30.843 [2024-09-29 16:47:52.298058] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.843 [2024-09-29 16:47:52.318685] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:30.843 BaseBdev1 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.843 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.843 [ 00:09:30.843 { 00:09:30.844 "name": "BaseBdev1", 00:09:30.844 "aliases": [ 00:09:30.844 "b29fb9c5-57b2-40e0-8c97-f5d3f149e880" 00:09:30.844 ], 00:09:30.844 "product_name": "Malloc disk", 00:09:30.844 "block_size": 512, 00:09:30.844 "num_blocks": 65536, 00:09:30.844 "uuid": "b29fb9c5-57b2-40e0-8c97-f5d3f149e880", 00:09:30.844 "assigned_rate_limits": { 00:09:30.844 "rw_ios_per_sec": 0, 00:09:30.844 "rw_mbytes_per_sec": 0, 00:09:30.844 "r_mbytes_per_sec": 0, 00:09:30.844 "w_mbytes_per_sec": 0 00:09:30.844 }, 00:09:30.844 "claimed": true, 00:09:30.844 "claim_type": "exclusive_write", 00:09:30.844 "zoned": false, 00:09:30.844 "supported_io_types": { 00:09:30.844 "read": true, 00:09:30.844 "write": true, 00:09:30.844 "unmap": true, 00:09:30.844 "flush": true, 00:09:30.844 "reset": true, 00:09:30.844 "nvme_admin": false, 00:09:30.844 "nvme_io": false, 00:09:30.844 "nvme_io_md": false, 00:09:30.844 "write_zeroes": true, 00:09:30.844 "zcopy": true, 00:09:30.844 "get_zone_info": false, 00:09:30.844 "zone_management": false, 00:09:30.844 "zone_append": false, 00:09:30.844 "compare": false, 00:09:30.844 "compare_and_write": false, 00:09:30.844 "abort": true, 00:09:30.844 "seek_hole": false, 00:09:30.844 "seek_data": false, 00:09:30.844 "copy": true, 00:09:30.844 "nvme_iov_md": false 00:09:30.844 }, 00:09:30.844 "memory_domains": [ 00:09:30.844 { 00:09:30.844 "dma_device_id": "system", 00:09:30.844 "dma_device_type": 1 00:09:30.844 }, 00:09:30.844 { 00:09:30.844 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.844 "dma_device_type": 2 00:09:30.844 } 00:09:30.844 ], 00:09:30.844 "driver_specific": {} 00:09:30.844 } 00:09:30.844 ] 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.844 "name": "Existed_Raid", 00:09:30.844 "uuid": "1c45a686-ea5d-48f6-ad85-7ec4e9f1785e", 00:09:30.844 "strip_size_kb": 64, 00:09:30.844 "state": "configuring", 00:09:30.844 "raid_level": "raid0", 00:09:30.844 "superblock": true, 00:09:30.844 "num_base_bdevs": 4, 00:09:30.844 "num_base_bdevs_discovered": 1, 00:09:30.844 "num_base_bdevs_operational": 4, 00:09:30.844 "base_bdevs_list": [ 00:09:30.844 { 00:09:30.844 "name": "BaseBdev1", 00:09:30.844 "uuid": "b29fb9c5-57b2-40e0-8c97-f5d3f149e880", 00:09:30.844 "is_configured": true, 00:09:30.844 "data_offset": 2048, 00:09:30.844 "data_size": 63488 00:09:30.844 }, 00:09:30.844 { 00:09:30.844 "name": "BaseBdev2", 00:09:30.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.844 "is_configured": false, 00:09:30.844 "data_offset": 0, 00:09:30.844 "data_size": 0 00:09:30.844 }, 00:09:30.844 { 00:09:30.844 "name": "BaseBdev3", 00:09:30.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.844 "is_configured": false, 00:09:30.844 "data_offset": 0, 00:09:30.844 "data_size": 0 00:09:30.844 }, 00:09:30.844 { 00:09:30.844 "name": "BaseBdev4", 00:09:30.844 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.844 "is_configured": false, 00:09:30.844 "data_offset": 0, 00:09:30.844 "data_size": 0 00:09:30.844 } 00:09:30.844 ] 00:09:30.844 }' 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.844 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.414 [2024-09-29 16:47:52.801872] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:31.414 [2024-09-29 16:47:52.801958] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.414 [2024-09-29 16:47:52.813910] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:31.414 [2024-09-29 16:47:52.815776] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:31.414 [2024-09-29 16:47:52.815847] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:31.414 [2024-09-29 16:47:52.815874] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:31.414 [2024-09-29 16:47:52.815895] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:31.414 [2024-09-29 16:47:52.815913] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:31.414 [2024-09-29 16:47:52.815933] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.414 "name": "Existed_Raid", 00:09:31.414 "uuid": "9e7df1fe-4350-4967-8108-9afdc7109328", 00:09:31.414 "strip_size_kb": 64, 00:09:31.414 "state": "configuring", 00:09:31.414 "raid_level": "raid0", 00:09:31.414 "superblock": true, 00:09:31.414 "num_base_bdevs": 4, 00:09:31.414 "num_base_bdevs_discovered": 1, 00:09:31.414 "num_base_bdevs_operational": 4, 00:09:31.414 "base_bdevs_list": [ 00:09:31.414 { 00:09:31.414 "name": "BaseBdev1", 00:09:31.414 "uuid": "b29fb9c5-57b2-40e0-8c97-f5d3f149e880", 00:09:31.414 "is_configured": true, 00:09:31.414 "data_offset": 2048, 00:09:31.414 "data_size": 63488 00:09:31.414 }, 00:09:31.414 { 00:09:31.414 "name": "BaseBdev2", 00:09:31.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.414 "is_configured": false, 00:09:31.414 "data_offset": 0, 00:09:31.414 "data_size": 0 00:09:31.414 }, 00:09:31.414 { 00:09:31.414 "name": "BaseBdev3", 00:09:31.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.414 "is_configured": false, 00:09:31.414 "data_offset": 0, 00:09:31.414 "data_size": 0 00:09:31.414 }, 00:09:31.414 { 00:09:31.414 "name": "BaseBdev4", 00:09:31.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.414 "is_configured": false, 00:09:31.414 "data_offset": 0, 00:09:31.414 "data_size": 0 00:09:31.414 } 00:09:31.414 ] 00:09:31.414 }' 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.414 16:47:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.674 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.675 [2024-09-29 16:47:53.309418] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:31.675 BaseBdev2 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.675 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.675 [ 00:09:31.675 { 00:09:31.675 "name": "BaseBdev2", 00:09:31.675 "aliases": [ 00:09:31.675 "e017fc4b-a3fe-40fe-b0da-7fc1dea542d7" 00:09:31.675 ], 00:09:31.675 "product_name": "Malloc disk", 00:09:31.675 "block_size": 512, 00:09:31.675 "num_blocks": 65536, 00:09:31.675 "uuid": "e017fc4b-a3fe-40fe-b0da-7fc1dea542d7", 00:09:31.675 "assigned_rate_limits": { 00:09:31.675 "rw_ios_per_sec": 0, 00:09:31.675 "rw_mbytes_per_sec": 0, 00:09:31.675 "r_mbytes_per_sec": 0, 00:09:31.675 "w_mbytes_per_sec": 0 00:09:31.675 }, 00:09:31.675 "claimed": true, 00:09:31.675 "claim_type": "exclusive_write", 00:09:31.675 "zoned": false, 00:09:31.675 "supported_io_types": { 00:09:31.675 "read": true, 00:09:31.675 "write": true, 00:09:31.675 "unmap": true, 00:09:31.675 "flush": true, 00:09:31.675 "reset": true, 00:09:31.675 "nvme_admin": false, 00:09:31.675 "nvme_io": false, 00:09:31.675 "nvme_io_md": false, 00:09:31.675 "write_zeroes": true, 00:09:31.675 "zcopy": true, 00:09:31.675 "get_zone_info": false, 00:09:31.675 "zone_management": false, 00:09:31.675 "zone_append": false, 00:09:31.675 "compare": false, 00:09:31.675 "compare_and_write": false, 00:09:31.675 "abort": true, 00:09:31.675 "seek_hole": false, 00:09:31.675 "seek_data": false, 00:09:31.675 "copy": true, 00:09:31.675 "nvme_iov_md": false 00:09:31.675 }, 00:09:31.675 "memory_domains": [ 00:09:31.675 { 00:09:31.675 "dma_device_id": "system", 00:09:31.675 "dma_device_type": 1 00:09:31.675 }, 00:09:31.675 { 00:09:31.675 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:31.675 "dma_device_type": 2 00:09:31.675 } 00:09:31.675 ], 00:09:31.675 "driver_specific": {} 00:09:31.675 } 00:09:31.675 ] 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.936 "name": "Existed_Raid", 00:09:31.936 "uuid": "9e7df1fe-4350-4967-8108-9afdc7109328", 00:09:31.936 "strip_size_kb": 64, 00:09:31.936 "state": "configuring", 00:09:31.936 "raid_level": "raid0", 00:09:31.936 "superblock": true, 00:09:31.936 "num_base_bdevs": 4, 00:09:31.936 "num_base_bdevs_discovered": 2, 00:09:31.936 "num_base_bdevs_operational": 4, 00:09:31.936 "base_bdevs_list": [ 00:09:31.936 { 00:09:31.936 "name": "BaseBdev1", 00:09:31.936 "uuid": "b29fb9c5-57b2-40e0-8c97-f5d3f149e880", 00:09:31.936 "is_configured": true, 00:09:31.936 "data_offset": 2048, 00:09:31.936 "data_size": 63488 00:09:31.936 }, 00:09:31.936 { 00:09:31.936 "name": "BaseBdev2", 00:09:31.936 "uuid": "e017fc4b-a3fe-40fe-b0da-7fc1dea542d7", 00:09:31.936 "is_configured": true, 00:09:31.936 "data_offset": 2048, 00:09:31.936 "data_size": 63488 00:09:31.936 }, 00:09:31.936 { 00:09:31.936 "name": "BaseBdev3", 00:09:31.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.936 "is_configured": false, 00:09:31.936 "data_offset": 0, 00:09:31.936 "data_size": 0 00:09:31.936 }, 00:09:31.936 { 00:09:31.936 "name": "BaseBdev4", 00:09:31.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.936 "is_configured": false, 00:09:31.936 "data_offset": 0, 00:09:31.936 "data_size": 0 00:09:31.936 } 00:09:31.936 ] 00:09:31.936 }' 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.936 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.195 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:32.195 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.195 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.196 [2024-09-29 16:47:53.839456] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:32.196 BaseBdev3 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.196 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.196 [ 00:09:32.196 { 00:09:32.196 "name": "BaseBdev3", 00:09:32.196 "aliases": [ 00:09:32.196 "d03e592d-d454-4205-b40f-647866ee3ae0" 00:09:32.196 ], 00:09:32.196 "product_name": "Malloc disk", 00:09:32.196 "block_size": 512, 00:09:32.196 "num_blocks": 65536, 00:09:32.196 "uuid": "d03e592d-d454-4205-b40f-647866ee3ae0", 00:09:32.196 "assigned_rate_limits": { 00:09:32.196 "rw_ios_per_sec": 0, 00:09:32.196 "rw_mbytes_per_sec": 0, 00:09:32.455 "r_mbytes_per_sec": 0, 00:09:32.455 "w_mbytes_per_sec": 0 00:09:32.455 }, 00:09:32.455 "claimed": true, 00:09:32.455 "claim_type": "exclusive_write", 00:09:32.455 "zoned": false, 00:09:32.455 "supported_io_types": { 00:09:32.455 "read": true, 00:09:32.455 "write": true, 00:09:32.455 "unmap": true, 00:09:32.455 "flush": true, 00:09:32.455 "reset": true, 00:09:32.455 "nvme_admin": false, 00:09:32.455 "nvme_io": false, 00:09:32.455 "nvme_io_md": false, 00:09:32.455 "write_zeroes": true, 00:09:32.455 "zcopy": true, 00:09:32.455 "get_zone_info": false, 00:09:32.455 "zone_management": false, 00:09:32.455 "zone_append": false, 00:09:32.455 "compare": false, 00:09:32.455 "compare_and_write": false, 00:09:32.455 "abort": true, 00:09:32.455 "seek_hole": false, 00:09:32.455 "seek_data": false, 00:09:32.455 "copy": true, 00:09:32.455 "nvme_iov_md": false 00:09:32.455 }, 00:09:32.455 "memory_domains": [ 00:09:32.455 { 00:09:32.455 "dma_device_id": "system", 00:09:32.455 "dma_device_type": 1 00:09:32.455 }, 00:09:32.455 { 00:09:32.455 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.455 "dma_device_type": 2 00:09:32.455 } 00:09:32.455 ], 00:09:32.455 "driver_specific": {} 00:09:32.455 } 00:09:32.455 ] 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.455 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.456 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.456 "name": "Existed_Raid", 00:09:32.456 "uuid": "9e7df1fe-4350-4967-8108-9afdc7109328", 00:09:32.456 "strip_size_kb": 64, 00:09:32.456 "state": "configuring", 00:09:32.456 "raid_level": "raid0", 00:09:32.456 "superblock": true, 00:09:32.456 "num_base_bdevs": 4, 00:09:32.456 "num_base_bdevs_discovered": 3, 00:09:32.456 "num_base_bdevs_operational": 4, 00:09:32.456 "base_bdevs_list": [ 00:09:32.456 { 00:09:32.456 "name": "BaseBdev1", 00:09:32.456 "uuid": "b29fb9c5-57b2-40e0-8c97-f5d3f149e880", 00:09:32.456 "is_configured": true, 00:09:32.456 "data_offset": 2048, 00:09:32.456 "data_size": 63488 00:09:32.456 }, 00:09:32.456 { 00:09:32.456 "name": "BaseBdev2", 00:09:32.456 "uuid": "e017fc4b-a3fe-40fe-b0da-7fc1dea542d7", 00:09:32.456 "is_configured": true, 00:09:32.456 "data_offset": 2048, 00:09:32.456 "data_size": 63488 00:09:32.456 }, 00:09:32.456 { 00:09:32.456 "name": "BaseBdev3", 00:09:32.456 "uuid": "d03e592d-d454-4205-b40f-647866ee3ae0", 00:09:32.456 "is_configured": true, 00:09:32.456 "data_offset": 2048, 00:09:32.456 "data_size": 63488 00:09:32.456 }, 00:09:32.456 { 00:09:32.456 "name": "BaseBdev4", 00:09:32.456 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.456 "is_configured": false, 00:09:32.456 "data_offset": 0, 00:09:32.456 "data_size": 0 00:09:32.456 } 00:09:32.456 ] 00:09:32.456 }' 00:09:32.456 16:47:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.456 16:47:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.716 [2024-09-29 16:47:54.329649] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:32.716 [2024-09-29 16:47:54.329955] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:32.716 [2024-09-29 16:47:54.330008] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:32.716 BaseBdev4 00:09:32.716 [2024-09-29 16:47:54.330312] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:32.716 [2024-09-29 16:47:54.330467] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:32.716 [2024-09-29 16:47:54.330529] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:32.716 [2024-09-29 16:47:54.330704] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.716 [ 00:09:32.716 { 00:09:32.716 "name": "BaseBdev4", 00:09:32.716 "aliases": [ 00:09:32.716 "a50aebea-cc3a-4ff4-b1b6-ad96ccb481ca" 00:09:32.716 ], 00:09:32.716 "product_name": "Malloc disk", 00:09:32.716 "block_size": 512, 00:09:32.716 "num_blocks": 65536, 00:09:32.716 "uuid": "a50aebea-cc3a-4ff4-b1b6-ad96ccb481ca", 00:09:32.716 "assigned_rate_limits": { 00:09:32.716 "rw_ios_per_sec": 0, 00:09:32.716 "rw_mbytes_per_sec": 0, 00:09:32.716 "r_mbytes_per_sec": 0, 00:09:32.716 "w_mbytes_per_sec": 0 00:09:32.716 }, 00:09:32.716 "claimed": true, 00:09:32.716 "claim_type": "exclusive_write", 00:09:32.716 "zoned": false, 00:09:32.716 "supported_io_types": { 00:09:32.716 "read": true, 00:09:32.716 "write": true, 00:09:32.716 "unmap": true, 00:09:32.716 "flush": true, 00:09:32.716 "reset": true, 00:09:32.716 "nvme_admin": false, 00:09:32.716 "nvme_io": false, 00:09:32.716 "nvme_io_md": false, 00:09:32.716 "write_zeroes": true, 00:09:32.716 "zcopy": true, 00:09:32.716 "get_zone_info": false, 00:09:32.716 "zone_management": false, 00:09:32.716 "zone_append": false, 00:09:32.716 "compare": false, 00:09:32.716 "compare_and_write": false, 00:09:32.716 "abort": true, 00:09:32.716 "seek_hole": false, 00:09:32.716 "seek_data": false, 00:09:32.716 "copy": true, 00:09:32.716 "nvme_iov_md": false 00:09:32.716 }, 00:09:32.716 "memory_domains": [ 00:09:32.716 { 00:09:32.716 "dma_device_id": "system", 00:09:32.716 "dma_device_type": 1 00:09:32.716 }, 00:09:32.716 { 00:09:32.716 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.716 "dma_device_type": 2 00:09:32.716 } 00:09:32.716 ], 00:09:32.716 "driver_specific": {} 00:09:32.716 } 00:09:32.716 ] 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.716 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.717 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.717 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:32.977 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.977 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.977 "name": "Existed_Raid", 00:09:32.977 "uuid": "9e7df1fe-4350-4967-8108-9afdc7109328", 00:09:32.977 "strip_size_kb": 64, 00:09:32.977 "state": "online", 00:09:32.977 "raid_level": "raid0", 00:09:32.977 "superblock": true, 00:09:32.977 "num_base_bdevs": 4, 00:09:32.977 "num_base_bdevs_discovered": 4, 00:09:32.977 "num_base_bdevs_operational": 4, 00:09:32.977 "base_bdevs_list": [ 00:09:32.977 { 00:09:32.977 "name": "BaseBdev1", 00:09:32.977 "uuid": "b29fb9c5-57b2-40e0-8c97-f5d3f149e880", 00:09:32.977 "is_configured": true, 00:09:32.977 "data_offset": 2048, 00:09:32.977 "data_size": 63488 00:09:32.977 }, 00:09:32.977 { 00:09:32.977 "name": "BaseBdev2", 00:09:32.977 "uuid": "e017fc4b-a3fe-40fe-b0da-7fc1dea542d7", 00:09:32.977 "is_configured": true, 00:09:32.977 "data_offset": 2048, 00:09:32.977 "data_size": 63488 00:09:32.977 }, 00:09:32.977 { 00:09:32.977 "name": "BaseBdev3", 00:09:32.977 "uuid": "d03e592d-d454-4205-b40f-647866ee3ae0", 00:09:32.977 "is_configured": true, 00:09:32.977 "data_offset": 2048, 00:09:32.977 "data_size": 63488 00:09:32.977 }, 00:09:32.977 { 00:09:32.977 "name": "BaseBdev4", 00:09:32.977 "uuid": "a50aebea-cc3a-4ff4-b1b6-ad96ccb481ca", 00:09:32.977 "is_configured": true, 00:09:32.977 "data_offset": 2048, 00:09:32.977 "data_size": 63488 00:09:32.977 } 00:09:32.977 ] 00:09:32.977 }' 00:09:32.977 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.977 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.328 [2024-09-29 16:47:54.813136] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:33.328 "name": "Existed_Raid", 00:09:33.328 "aliases": [ 00:09:33.328 "9e7df1fe-4350-4967-8108-9afdc7109328" 00:09:33.328 ], 00:09:33.328 "product_name": "Raid Volume", 00:09:33.328 "block_size": 512, 00:09:33.328 "num_blocks": 253952, 00:09:33.328 "uuid": "9e7df1fe-4350-4967-8108-9afdc7109328", 00:09:33.328 "assigned_rate_limits": { 00:09:33.328 "rw_ios_per_sec": 0, 00:09:33.328 "rw_mbytes_per_sec": 0, 00:09:33.328 "r_mbytes_per_sec": 0, 00:09:33.328 "w_mbytes_per_sec": 0 00:09:33.328 }, 00:09:33.328 "claimed": false, 00:09:33.328 "zoned": false, 00:09:33.328 "supported_io_types": { 00:09:33.328 "read": true, 00:09:33.328 "write": true, 00:09:33.328 "unmap": true, 00:09:33.328 "flush": true, 00:09:33.328 "reset": true, 00:09:33.328 "nvme_admin": false, 00:09:33.328 "nvme_io": false, 00:09:33.328 "nvme_io_md": false, 00:09:33.328 "write_zeroes": true, 00:09:33.328 "zcopy": false, 00:09:33.328 "get_zone_info": false, 00:09:33.328 "zone_management": false, 00:09:33.328 "zone_append": false, 00:09:33.328 "compare": false, 00:09:33.328 "compare_and_write": false, 00:09:33.328 "abort": false, 00:09:33.328 "seek_hole": false, 00:09:33.328 "seek_data": false, 00:09:33.328 "copy": false, 00:09:33.328 "nvme_iov_md": false 00:09:33.328 }, 00:09:33.328 "memory_domains": [ 00:09:33.328 { 00:09:33.328 "dma_device_id": "system", 00:09:33.328 "dma_device_type": 1 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.328 "dma_device_type": 2 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "dma_device_id": "system", 00:09:33.328 "dma_device_type": 1 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.328 "dma_device_type": 2 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "dma_device_id": "system", 00:09:33.328 "dma_device_type": 1 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.328 "dma_device_type": 2 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "dma_device_id": "system", 00:09:33.328 "dma_device_type": 1 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.328 "dma_device_type": 2 00:09:33.328 } 00:09:33.328 ], 00:09:33.328 "driver_specific": { 00:09:33.328 "raid": { 00:09:33.328 "uuid": "9e7df1fe-4350-4967-8108-9afdc7109328", 00:09:33.328 "strip_size_kb": 64, 00:09:33.328 "state": "online", 00:09:33.328 "raid_level": "raid0", 00:09:33.328 "superblock": true, 00:09:33.328 "num_base_bdevs": 4, 00:09:33.328 "num_base_bdevs_discovered": 4, 00:09:33.328 "num_base_bdevs_operational": 4, 00:09:33.328 "base_bdevs_list": [ 00:09:33.328 { 00:09:33.328 "name": "BaseBdev1", 00:09:33.328 "uuid": "b29fb9c5-57b2-40e0-8c97-f5d3f149e880", 00:09:33.328 "is_configured": true, 00:09:33.328 "data_offset": 2048, 00:09:33.328 "data_size": 63488 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "name": "BaseBdev2", 00:09:33.328 "uuid": "e017fc4b-a3fe-40fe-b0da-7fc1dea542d7", 00:09:33.328 "is_configured": true, 00:09:33.328 "data_offset": 2048, 00:09:33.328 "data_size": 63488 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "name": "BaseBdev3", 00:09:33.328 "uuid": "d03e592d-d454-4205-b40f-647866ee3ae0", 00:09:33.328 "is_configured": true, 00:09:33.328 "data_offset": 2048, 00:09:33.328 "data_size": 63488 00:09:33.328 }, 00:09:33.328 { 00:09:33.328 "name": "BaseBdev4", 00:09:33.328 "uuid": "a50aebea-cc3a-4ff4-b1b6-ad96ccb481ca", 00:09:33.328 "is_configured": true, 00:09:33.328 "data_offset": 2048, 00:09:33.328 "data_size": 63488 00:09:33.328 } 00:09:33.328 ] 00:09:33.328 } 00:09:33.328 } 00:09:33.328 }' 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:33.328 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:33.328 BaseBdev2 00:09:33.329 BaseBdev3 00:09:33.329 BaseBdev4' 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.329 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.622 16:47:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.622 [2024-09-29 16:47:55.100423] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:33.622 [2024-09-29 16:47:55.100497] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:33.622 [2024-09-29 16:47:55.100567] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:33.622 "name": "Existed_Raid", 00:09:33.622 "uuid": "9e7df1fe-4350-4967-8108-9afdc7109328", 00:09:33.622 "strip_size_kb": 64, 00:09:33.622 "state": "offline", 00:09:33.622 "raid_level": "raid0", 00:09:33.622 "superblock": true, 00:09:33.622 "num_base_bdevs": 4, 00:09:33.622 "num_base_bdevs_discovered": 3, 00:09:33.622 "num_base_bdevs_operational": 3, 00:09:33.622 "base_bdevs_list": [ 00:09:33.622 { 00:09:33.622 "name": null, 00:09:33.622 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:33.622 "is_configured": false, 00:09:33.622 "data_offset": 0, 00:09:33.622 "data_size": 63488 00:09:33.622 }, 00:09:33.622 { 00:09:33.622 "name": "BaseBdev2", 00:09:33.622 "uuid": "e017fc4b-a3fe-40fe-b0da-7fc1dea542d7", 00:09:33.622 "is_configured": true, 00:09:33.622 "data_offset": 2048, 00:09:33.622 "data_size": 63488 00:09:33.622 }, 00:09:33.622 { 00:09:33.622 "name": "BaseBdev3", 00:09:33.622 "uuid": "d03e592d-d454-4205-b40f-647866ee3ae0", 00:09:33.622 "is_configured": true, 00:09:33.622 "data_offset": 2048, 00:09:33.622 "data_size": 63488 00:09:33.622 }, 00:09:33.622 { 00:09:33.622 "name": "BaseBdev4", 00:09:33.622 "uuid": "a50aebea-cc3a-4ff4-b1b6-ad96ccb481ca", 00:09:33.622 "is_configured": true, 00:09:33.622 "data_offset": 2048, 00:09:33.622 "data_size": 63488 00:09:33.622 } 00:09:33.622 ] 00:09:33.622 }' 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:33.622 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 [2024-09-29 16:47:55.614969] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 [2024-09-29 16:47:55.686016] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 [2024-09-29 16:47:55.756899] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:34.192 [2024-09-29 16:47:55.756945] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.192 BaseBdev2 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.192 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.193 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.193 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:34.193 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.193 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.193 [ 00:09:34.193 { 00:09:34.193 "name": "BaseBdev2", 00:09:34.193 "aliases": [ 00:09:34.193 "0be71bcb-b765-42d5-be63-348e64ca28da" 00:09:34.193 ], 00:09:34.193 "product_name": "Malloc disk", 00:09:34.193 "block_size": 512, 00:09:34.193 "num_blocks": 65536, 00:09:34.193 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:34.193 "assigned_rate_limits": { 00:09:34.193 "rw_ios_per_sec": 0, 00:09:34.193 "rw_mbytes_per_sec": 0, 00:09:34.193 "r_mbytes_per_sec": 0, 00:09:34.193 "w_mbytes_per_sec": 0 00:09:34.193 }, 00:09:34.193 "claimed": false, 00:09:34.193 "zoned": false, 00:09:34.193 "supported_io_types": { 00:09:34.193 "read": true, 00:09:34.452 "write": true, 00:09:34.452 "unmap": true, 00:09:34.452 "flush": true, 00:09:34.452 "reset": true, 00:09:34.452 "nvme_admin": false, 00:09:34.452 "nvme_io": false, 00:09:34.452 "nvme_io_md": false, 00:09:34.452 "write_zeroes": true, 00:09:34.452 "zcopy": true, 00:09:34.452 "get_zone_info": false, 00:09:34.452 "zone_management": false, 00:09:34.452 "zone_append": false, 00:09:34.452 "compare": false, 00:09:34.452 "compare_and_write": false, 00:09:34.452 "abort": true, 00:09:34.452 "seek_hole": false, 00:09:34.452 "seek_data": false, 00:09:34.452 "copy": true, 00:09:34.452 "nvme_iov_md": false 00:09:34.452 }, 00:09:34.452 "memory_domains": [ 00:09:34.452 { 00:09:34.452 "dma_device_id": "system", 00:09:34.452 "dma_device_type": 1 00:09:34.452 }, 00:09:34.452 { 00:09:34.452 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.452 "dma_device_type": 2 00:09:34.452 } 00:09:34.452 ], 00:09:34.452 "driver_specific": {} 00:09:34.452 } 00:09:34.452 ] 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.452 BaseBdev3 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.452 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.453 [ 00:09:34.453 { 00:09:34.453 "name": "BaseBdev3", 00:09:34.453 "aliases": [ 00:09:34.453 "97c352a4-b8c3-41c3-bdaf-be42dc952efc" 00:09:34.453 ], 00:09:34.453 "product_name": "Malloc disk", 00:09:34.453 "block_size": 512, 00:09:34.453 "num_blocks": 65536, 00:09:34.453 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:34.453 "assigned_rate_limits": { 00:09:34.453 "rw_ios_per_sec": 0, 00:09:34.453 "rw_mbytes_per_sec": 0, 00:09:34.453 "r_mbytes_per_sec": 0, 00:09:34.453 "w_mbytes_per_sec": 0 00:09:34.453 }, 00:09:34.453 "claimed": false, 00:09:34.453 "zoned": false, 00:09:34.453 "supported_io_types": { 00:09:34.453 "read": true, 00:09:34.453 "write": true, 00:09:34.453 "unmap": true, 00:09:34.453 "flush": true, 00:09:34.453 "reset": true, 00:09:34.453 "nvme_admin": false, 00:09:34.453 "nvme_io": false, 00:09:34.453 "nvme_io_md": false, 00:09:34.453 "write_zeroes": true, 00:09:34.453 "zcopy": true, 00:09:34.453 "get_zone_info": false, 00:09:34.453 "zone_management": false, 00:09:34.453 "zone_append": false, 00:09:34.453 "compare": false, 00:09:34.453 "compare_and_write": false, 00:09:34.453 "abort": true, 00:09:34.453 "seek_hole": false, 00:09:34.453 "seek_data": false, 00:09:34.453 "copy": true, 00:09:34.453 "nvme_iov_md": false 00:09:34.453 }, 00:09:34.453 "memory_domains": [ 00:09:34.453 { 00:09:34.453 "dma_device_id": "system", 00:09:34.453 "dma_device_type": 1 00:09:34.453 }, 00:09:34.453 { 00:09:34.453 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.453 "dma_device_type": 2 00:09:34.453 } 00:09:34.453 ], 00:09:34.453 "driver_specific": {} 00:09:34.453 } 00:09:34.453 ] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.453 BaseBdev4 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.453 [ 00:09:34.453 { 00:09:34.453 "name": "BaseBdev4", 00:09:34.453 "aliases": [ 00:09:34.453 "ca5447e8-2e17-43a5-88cf-54d20a67b7bf" 00:09:34.453 ], 00:09:34.453 "product_name": "Malloc disk", 00:09:34.453 "block_size": 512, 00:09:34.453 "num_blocks": 65536, 00:09:34.453 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:34.453 "assigned_rate_limits": { 00:09:34.453 "rw_ios_per_sec": 0, 00:09:34.453 "rw_mbytes_per_sec": 0, 00:09:34.453 "r_mbytes_per_sec": 0, 00:09:34.453 "w_mbytes_per_sec": 0 00:09:34.453 }, 00:09:34.453 "claimed": false, 00:09:34.453 "zoned": false, 00:09:34.453 "supported_io_types": { 00:09:34.453 "read": true, 00:09:34.453 "write": true, 00:09:34.453 "unmap": true, 00:09:34.453 "flush": true, 00:09:34.453 "reset": true, 00:09:34.453 "nvme_admin": false, 00:09:34.453 "nvme_io": false, 00:09:34.453 "nvme_io_md": false, 00:09:34.453 "write_zeroes": true, 00:09:34.453 "zcopy": true, 00:09:34.453 "get_zone_info": false, 00:09:34.453 "zone_management": false, 00:09:34.453 "zone_append": false, 00:09:34.453 "compare": false, 00:09:34.453 "compare_and_write": false, 00:09:34.453 "abort": true, 00:09:34.453 "seek_hole": false, 00:09:34.453 "seek_data": false, 00:09:34.453 "copy": true, 00:09:34.453 "nvme_iov_md": false 00:09:34.453 }, 00:09:34.453 "memory_domains": [ 00:09:34.453 { 00:09:34.453 "dma_device_id": "system", 00:09:34.453 "dma_device_type": 1 00:09:34.453 }, 00:09:34.453 { 00:09:34.453 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:34.453 "dma_device_type": 2 00:09:34.453 } 00:09:34.453 ], 00:09:34.453 "driver_specific": {} 00:09:34.453 } 00:09:34.453 ] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.453 [2024-09-29 16:47:55.984286] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:34.453 [2024-09-29 16:47:55.984395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:34.453 [2024-09-29 16:47:55.984436] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:34.453 [2024-09-29 16:47:55.986250] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:34.453 [2024-09-29 16:47:55.986350] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:34.453 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.454 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.454 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.454 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.454 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.454 16:47:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.454 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.454 16:47:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.454 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.454 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.454 "name": "Existed_Raid", 00:09:34.454 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:34.454 "strip_size_kb": 64, 00:09:34.454 "state": "configuring", 00:09:34.454 "raid_level": "raid0", 00:09:34.454 "superblock": true, 00:09:34.454 "num_base_bdevs": 4, 00:09:34.454 "num_base_bdevs_discovered": 3, 00:09:34.454 "num_base_bdevs_operational": 4, 00:09:34.454 "base_bdevs_list": [ 00:09:34.454 { 00:09:34.454 "name": "BaseBdev1", 00:09:34.454 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.454 "is_configured": false, 00:09:34.454 "data_offset": 0, 00:09:34.454 "data_size": 0 00:09:34.454 }, 00:09:34.454 { 00:09:34.454 "name": "BaseBdev2", 00:09:34.454 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:34.454 "is_configured": true, 00:09:34.454 "data_offset": 2048, 00:09:34.454 "data_size": 63488 00:09:34.454 }, 00:09:34.454 { 00:09:34.454 "name": "BaseBdev3", 00:09:34.454 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:34.454 "is_configured": true, 00:09:34.454 "data_offset": 2048, 00:09:34.454 "data_size": 63488 00:09:34.454 }, 00:09:34.454 { 00:09:34.454 "name": "BaseBdev4", 00:09:34.454 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:34.454 "is_configured": true, 00:09:34.454 "data_offset": 2048, 00:09:34.454 "data_size": 63488 00:09:34.454 } 00:09:34.454 ] 00:09:34.454 }' 00:09:34.454 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.454 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.022 [2024-09-29 16:47:56.435498] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:35.022 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.023 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.023 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.023 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.023 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.023 "name": "Existed_Raid", 00:09:35.023 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:35.023 "strip_size_kb": 64, 00:09:35.023 "state": "configuring", 00:09:35.023 "raid_level": "raid0", 00:09:35.023 "superblock": true, 00:09:35.023 "num_base_bdevs": 4, 00:09:35.023 "num_base_bdevs_discovered": 2, 00:09:35.023 "num_base_bdevs_operational": 4, 00:09:35.023 "base_bdevs_list": [ 00:09:35.023 { 00:09:35.023 "name": "BaseBdev1", 00:09:35.023 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:35.023 "is_configured": false, 00:09:35.023 "data_offset": 0, 00:09:35.023 "data_size": 0 00:09:35.023 }, 00:09:35.023 { 00:09:35.023 "name": null, 00:09:35.023 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:35.023 "is_configured": false, 00:09:35.023 "data_offset": 0, 00:09:35.023 "data_size": 63488 00:09:35.023 }, 00:09:35.023 { 00:09:35.023 "name": "BaseBdev3", 00:09:35.023 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:35.023 "is_configured": true, 00:09:35.023 "data_offset": 2048, 00:09:35.023 "data_size": 63488 00:09:35.023 }, 00:09:35.023 { 00:09:35.023 "name": "BaseBdev4", 00:09:35.023 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:35.023 "is_configured": true, 00:09:35.023 "data_offset": 2048, 00:09:35.023 "data_size": 63488 00:09:35.023 } 00:09:35.023 ] 00:09:35.023 }' 00:09:35.023 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.023 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.282 [2024-09-29 16:47:56.937520] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:35.282 BaseBdev1 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.282 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.541 [ 00:09:35.541 { 00:09:35.541 "name": "BaseBdev1", 00:09:35.541 "aliases": [ 00:09:35.541 "787dc674-5aa9-4365-a6a6-bcdbd83555a8" 00:09:35.541 ], 00:09:35.541 "product_name": "Malloc disk", 00:09:35.541 "block_size": 512, 00:09:35.541 "num_blocks": 65536, 00:09:35.541 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:35.541 "assigned_rate_limits": { 00:09:35.541 "rw_ios_per_sec": 0, 00:09:35.541 "rw_mbytes_per_sec": 0, 00:09:35.541 "r_mbytes_per_sec": 0, 00:09:35.541 "w_mbytes_per_sec": 0 00:09:35.541 }, 00:09:35.541 "claimed": true, 00:09:35.541 "claim_type": "exclusive_write", 00:09:35.541 "zoned": false, 00:09:35.541 "supported_io_types": { 00:09:35.541 "read": true, 00:09:35.541 "write": true, 00:09:35.541 "unmap": true, 00:09:35.541 "flush": true, 00:09:35.541 "reset": true, 00:09:35.541 "nvme_admin": false, 00:09:35.541 "nvme_io": false, 00:09:35.541 "nvme_io_md": false, 00:09:35.541 "write_zeroes": true, 00:09:35.541 "zcopy": true, 00:09:35.541 "get_zone_info": false, 00:09:35.541 "zone_management": false, 00:09:35.541 "zone_append": false, 00:09:35.541 "compare": false, 00:09:35.541 "compare_and_write": false, 00:09:35.541 "abort": true, 00:09:35.541 "seek_hole": false, 00:09:35.541 "seek_data": false, 00:09:35.541 "copy": true, 00:09:35.541 "nvme_iov_md": false 00:09:35.541 }, 00:09:35.541 "memory_domains": [ 00:09:35.541 { 00:09:35.541 "dma_device_id": "system", 00:09:35.541 "dma_device_type": 1 00:09:35.541 }, 00:09:35.541 { 00:09:35.541 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.541 "dma_device_type": 2 00:09:35.541 } 00:09:35.541 ], 00:09:35.541 "driver_specific": {} 00:09:35.541 } 00:09:35.541 ] 00:09:35.541 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.541 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:35.541 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:35.541 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.542 16:47:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.542 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.542 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.542 "name": "Existed_Raid", 00:09:35.542 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:35.542 "strip_size_kb": 64, 00:09:35.542 "state": "configuring", 00:09:35.542 "raid_level": "raid0", 00:09:35.542 "superblock": true, 00:09:35.542 "num_base_bdevs": 4, 00:09:35.542 "num_base_bdevs_discovered": 3, 00:09:35.542 "num_base_bdevs_operational": 4, 00:09:35.542 "base_bdevs_list": [ 00:09:35.542 { 00:09:35.542 "name": "BaseBdev1", 00:09:35.542 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:35.542 "is_configured": true, 00:09:35.542 "data_offset": 2048, 00:09:35.542 "data_size": 63488 00:09:35.542 }, 00:09:35.542 { 00:09:35.542 "name": null, 00:09:35.542 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:35.542 "is_configured": false, 00:09:35.542 "data_offset": 0, 00:09:35.542 "data_size": 63488 00:09:35.542 }, 00:09:35.542 { 00:09:35.542 "name": "BaseBdev3", 00:09:35.542 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:35.542 "is_configured": true, 00:09:35.542 "data_offset": 2048, 00:09:35.542 "data_size": 63488 00:09:35.542 }, 00:09:35.542 { 00:09:35.542 "name": "BaseBdev4", 00:09:35.542 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:35.542 "is_configured": true, 00:09:35.542 "data_offset": 2048, 00:09:35.542 "data_size": 63488 00:09:35.542 } 00:09:35.542 ] 00:09:35.542 }' 00:09:35.542 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.542 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.801 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.801 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:35.801 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.801 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.801 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.060 [2024-09-29 16:47:57.484617] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.060 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.061 "name": "Existed_Raid", 00:09:36.061 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:36.061 "strip_size_kb": 64, 00:09:36.061 "state": "configuring", 00:09:36.061 "raid_level": "raid0", 00:09:36.061 "superblock": true, 00:09:36.061 "num_base_bdevs": 4, 00:09:36.061 "num_base_bdevs_discovered": 2, 00:09:36.061 "num_base_bdevs_operational": 4, 00:09:36.061 "base_bdevs_list": [ 00:09:36.061 { 00:09:36.061 "name": "BaseBdev1", 00:09:36.061 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:36.061 "is_configured": true, 00:09:36.061 "data_offset": 2048, 00:09:36.061 "data_size": 63488 00:09:36.061 }, 00:09:36.061 { 00:09:36.061 "name": null, 00:09:36.061 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:36.061 "is_configured": false, 00:09:36.061 "data_offset": 0, 00:09:36.061 "data_size": 63488 00:09:36.061 }, 00:09:36.061 { 00:09:36.061 "name": null, 00:09:36.061 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:36.061 "is_configured": false, 00:09:36.061 "data_offset": 0, 00:09:36.061 "data_size": 63488 00:09:36.061 }, 00:09:36.061 { 00:09:36.061 "name": "BaseBdev4", 00:09:36.061 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:36.061 "is_configured": true, 00:09:36.061 "data_offset": 2048, 00:09:36.061 "data_size": 63488 00:09:36.061 } 00:09:36.061 ] 00:09:36.061 }' 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.061 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.320 [2024-09-29 16:47:57.943875] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.320 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.320 "name": "Existed_Raid", 00:09:36.320 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:36.320 "strip_size_kb": 64, 00:09:36.320 "state": "configuring", 00:09:36.320 "raid_level": "raid0", 00:09:36.320 "superblock": true, 00:09:36.320 "num_base_bdevs": 4, 00:09:36.320 "num_base_bdevs_discovered": 3, 00:09:36.320 "num_base_bdevs_operational": 4, 00:09:36.320 "base_bdevs_list": [ 00:09:36.320 { 00:09:36.320 "name": "BaseBdev1", 00:09:36.320 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:36.320 "is_configured": true, 00:09:36.320 "data_offset": 2048, 00:09:36.320 "data_size": 63488 00:09:36.320 }, 00:09:36.320 { 00:09:36.320 "name": null, 00:09:36.320 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:36.320 "is_configured": false, 00:09:36.320 "data_offset": 0, 00:09:36.320 "data_size": 63488 00:09:36.320 }, 00:09:36.320 { 00:09:36.320 "name": "BaseBdev3", 00:09:36.320 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:36.320 "is_configured": true, 00:09:36.320 "data_offset": 2048, 00:09:36.320 "data_size": 63488 00:09:36.320 }, 00:09:36.320 { 00:09:36.320 "name": "BaseBdev4", 00:09:36.320 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:36.321 "is_configured": true, 00:09:36.321 "data_offset": 2048, 00:09:36.321 "data_size": 63488 00:09:36.321 } 00:09:36.321 ] 00:09:36.321 }' 00:09:36.321 16:47:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.321 16:47:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.889 [2024-09-29 16:47:58.415050] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.889 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.890 "name": "Existed_Raid", 00:09:36.890 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:36.890 "strip_size_kb": 64, 00:09:36.890 "state": "configuring", 00:09:36.890 "raid_level": "raid0", 00:09:36.890 "superblock": true, 00:09:36.890 "num_base_bdevs": 4, 00:09:36.890 "num_base_bdevs_discovered": 2, 00:09:36.890 "num_base_bdevs_operational": 4, 00:09:36.890 "base_bdevs_list": [ 00:09:36.890 { 00:09:36.890 "name": null, 00:09:36.890 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:36.890 "is_configured": false, 00:09:36.890 "data_offset": 0, 00:09:36.890 "data_size": 63488 00:09:36.890 }, 00:09:36.890 { 00:09:36.890 "name": null, 00:09:36.890 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:36.890 "is_configured": false, 00:09:36.890 "data_offset": 0, 00:09:36.890 "data_size": 63488 00:09:36.890 }, 00:09:36.890 { 00:09:36.890 "name": "BaseBdev3", 00:09:36.890 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:36.890 "is_configured": true, 00:09:36.890 "data_offset": 2048, 00:09:36.890 "data_size": 63488 00:09:36.890 }, 00:09:36.890 { 00:09:36.890 "name": "BaseBdev4", 00:09:36.890 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:36.890 "is_configured": true, 00:09:36.890 "data_offset": 2048, 00:09:36.890 "data_size": 63488 00:09:36.890 } 00:09:36.890 ] 00:09:36.890 }' 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.890 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.458 [2024-09-29 16:47:58.888715] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.458 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.458 "name": "Existed_Raid", 00:09:37.458 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:37.458 "strip_size_kb": 64, 00:09:37.459 "state": "configuring", 00:09:37.459 "raid_level": "raid0", 00:09:37.459 "superblock": true, 00:09:37.459 "num_base_bdevs": 4, 00:09:37.459 "num_base_bdevs_discovered": 3, 00:09:37.459 "num_base_bdevs_operational": 4, 00:09:37.459 "base_bdevs_list": [ 00:09:37.459 { 00:09:37.459 "name": null, 00:09:37.459 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:37.459 "is_configured": false, 00:09:37.459 "data_offset": 0, 00:09:37.459 "data_size": 63488 00:09:37.459 }, 00:09:37.459 { 00:09:37.459 "name": "BaseBdev2", 00:09:37.459 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:37.459 "is_configured": true, 00:09:37.459 "data_offset": 2048, 00:09:37.459 "data_size": 63488 00:09:37.459 }, 00:09:37.459 { 00:09:37.459 "name": "BaseBdev3", 00:09:37.459 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:37.459 "is_configured": true, 00:09:37.459 "data_offset": 2048, 00:09:37.459 "data_size": 63488 00:09:37.459 }, 00:09:37.459 { 00:09:37.459 "name": "BaseBdev4", 00:09:37.459 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:37.459 "is_configured": true, 00:09:37.459 "data_offset": 2048, 00:09:37.459 "data_size": 63488 00:09:37.459 } 00:09:37.459 ] 00:09:37.459 }' 00:09:37.459 16:47:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.459 16:47:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:37.718 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.977 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 787dc674-5aa9-4365-a6a6-bcdbd83555a8 00:09:37.977 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.977 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.977 [2024-09-29 16:47:59.426725] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:37.978 [2024-09-29 16:47:59.427010] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:37.978 [2024-09-29 16:47:59.427076] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:37.978 [2024-09-29 16:47:59.427369] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:37.978 NewBaseBdev 00:09:37.978 [2024-09-29 16:47:59.427528] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:37.978 [2024-09-29 16:47:59.427549] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:37.978 [2024-09-29 16:47:59.427644] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.978 [ 00:09:37.978 { 00:09:37.978 "name": "NewBaseBdev", 00:09:37.978 "aliases": [ 00:09:37.978 "787dc674-5aa9-4365-a6a6-bcdbd83555a8" 00:09:37.978 ], 00:09:37.978 "product_name": "Malloc disk", 00:09:37.978 "block_size": 512, 00:09:37.978 "num_blocks": 65536, 00:09:37.978 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:37.978 "assigned_rate_limits": { 00:09:37.978 "rw_ios_per_sec": 0, 00:09:37.978 "rw_mbytes_per_sec": 0, 00:09:37.978 "r_mbytes_per_sec": 0, 00:09:37.978 "w_mbytes_per_sec": 0 00:09:37.978 }, 00:09:37.978 "claimed": true, 00:09:37.978 "claim_type": "exclusive_write", 00:09:37.978 "zoned": false, 00:09:37.978 "supported_io_types": { 00:09:37.978 "read": true, 00:09:37.978 "write": true, 00:09:37.978 "unmap": true, 00:09:37.978 "flush": true, 00:09:37.978 "reset": true, 00:09:37.978 "nvme_admin": false, 00:09:37.978 "nvme_io": false, 00:09:37.978 "nvme_io_md": false, 00:09:37.978 "write_zeroes": true, 00:09:37.978 "zcopy": true, 00:09:37.978 "get_zone_info": false, 00:09:37.978 "zone_management": false, 00:09:37.978 "zone_append": false, 00:09:37.978 "compare": false, 00:09:37.978 "compare_and_write": false, 00:09:37.978 "abort": true, 00:09:37.978 "seek_hole": false, 00:09:37.978 "seek_data": false, 00:09:37.978 "copy": true, 00:09:37.978 "nvme_iov_md": false 00:09:37.978 }, 00:09:37.978 "memory_domains": [ 00:09:37.978 { 00:09:37.978 "dma_device_id": "system", 00:09:37.978 "dma_device_type": 1 00:09:37.978 }, 00:09:37.978 { 00:09:37.978 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.978 "dma_device_type": 2 00:09:37.978 } 00:09:37.978 ], 00:09:37.978 "driver_specific": {} 00:09:37.978 } 00:09:37.978 ] 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.978 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.978 "name": "Existed_Raid", 00:09:37.978 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:37.978 "strip_size_kb": 64, 00:09:37.978 "state": "online", 00:09:37.978 "raid_level": "raid0", 00:09:37.978 "superblock": true, 00:09:37.978 "num_base_bdevs": 4, 00:09:37.978 "num_base_bdevs_discovered": 4, 00:09:37.978 "num_base_bdevs_operational": 4, 00:09:37.978 "base_bdevs_list": [ 00:09:37.978 { 00:09:37.978 "name": "NewBaseBdev", 00:09:37.978 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:37.978 "is_configured": true, 00:09:37.978 "data_offset": 2048, 00:09:37.978 "data_size": 63488 00:09:37.978 }, 00:09:37.978 { 00:09:37.978 "name": "BaseBdev2", 00:09:37.978 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:37.978 "is_configured": true, 00:09:37.978 "data_offset": 2048, 00:09:37.978 "data_size": 63488 00:09:37.979 }, 00:09:37.979 { 00:09:37.979 "name": "BaseBdev3", 00:09:37.979 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:37.979 "is_configured": true, 00:09:37.979 "data_offset": 2048, 00:09:37.979 "data_size": 63488 00:09:37.979 }, 00:09:37.979 { 00:09:37.979 "name": "BaseBdev4", 00:09:37.979 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:37.979 "is_configured": true, 00:09:37.979 "data_offset": 2048, 00:09:37.979 "data_size": 63488 00:09:37.979 } 00:09:37.979 ] 00:09:37.979 }' 00:09:37.979 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.979 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.547 [2024-09-29 16:47:59.942167] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.547 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:38.547 "name": "Existed_Raid", 00:09:38.547 "aliases": [ 00:09:38.547 "d50b70c6-cbc8-4a4a-b49f-3be5429fa296" 00:09:38.547 ], 00:09:38.547 "product_name": "Raid Volume", 00:09:38.547 "block_size": 512, 00:09:38.547 "num_blocks": 253952, 00:09:38.547 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:38.547 "assigned_rate_limits": { 00:09:38.547 "rw_ios_per_sec": 0, 00:09:38.547 "rw_mbytes_per_sec": 0, 00:09:38.547 "r_mbytes_per_sec": 0, 00:09:38.547 "w_mbytes_per_sec": 0 00:09:38.547 }, 00:09:38.547 "claimed": false, 00:09:38.547 "zoned": false, 00:09:38.547 "supported_io_types": { 00:09:38.547 "read": true, 00:09:38.547 "write": true, 00:09:38.547 "unmap": true, 00:09:38.547 "flush": true, 00:09:38.547 "reset": true, 00:09:38.547 "nvme_admin": false, 00:09:38.547 "nvme_io": false, 00:09:38.547 "nvme_io_md": false, 00:09:38.547 "write_zeroes": true, 00:09:38.547 "zcopy": false, 00:09:38.547 "get_zone_info": false, 00:09:38.547 "zone_management": false, 00:09:38.547 "zone_append": false, 00:09:38.547 "compare": false, 00:09:38.547 "compare_and_write": false, 00:09:38.547 "abort": false, 00:09:38.547 "seek_hole": false, 00:09:38.547 "seek_data": false, 00:09:38.547 "copy": false, 00:09:38.547 "nvme_iov_md": false 00:09:38.547 }, 00:09:38.547 "memory_domains": [ 00:09:38.547 { 00:09:38.547 "dma_device_id": "system", 00:09:38.547 "dma_device_type": 1 00:09:38.547 }, 00:09:38.547 { 00:09:38.547 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.547 "dma_device_type": 2 00:09:38.547 }, 00:09:38.547 { 00:09:38.547 "dma_device_id": "system", 00:09:38.547 "dma_device_type": 1 00:09:38.547 }, 00:09:38.547 { 00:09:38.547 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.547 "dma_device_type": 2 00:09:38.547 }, 00:09:38.547 { 00:09:38.548 "dma_device_id": "system", 00:09:38.548 "dma_device_type": 1 00:09:38.548 }, 00:09:38.548 { 00:09:38.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.548 "dma_device_type": 2 00:09:38.548 }, 00:09:38.548 { 00:09:38.548 "dma_device_id": "system", 00:09:38.548 "dma_device_type": 1 00:09:38.548 }, 00:09:38.548 { 00:09:38.548 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.548 "dma_device_type": 2 00:09:38.548 } 00:09:38.548 ], 00:09:38.548 "driver_specific": { 00:09:38.548 "raid": { 00:09:38.548 "uuid": "d50b70c6-cbc8-4a4a-b49f-3be5429fa296", 00:09:38.548 "strip_size_kb": 64, 00:09:38.548 "state": "online", 00:09:38.548 "raid_level": "raid0", 00:09:38.548 "superblock": true, 00:09:38.548 "num_base_bdevs": 4, 00:09:38.548 "num_base_bdevs_discovered": 4, 00:09:38.548 "num_base_bdevs_operational": 4, 00:09:38.548 "base_bdevs_list": [ 00:09:38.548 { 00:09:38.548 "name": "NewBaseBdev", 00:09:38.548 "uuid": "787dc674-5aa9-4365-a6a6-bcdbd83555a8", 00:09:38.548 "is_configured": true, 00:09:38.548 "data_offset": 2048, 00:09:38.548 "data_size": 63488 00:09:38.548 }, 00:09:38.548 { 00:09:38.548 "name": "BaseBdev2", 00:09:38.548 "uuid": "0be71bcb-b765-42d5-be63-348e64ca28da", 00:09:38.548 "is_configured": true, 00:09:38.548 "data_offset": 2048, 00:09:38.548 "data_size": 63488 00:09:38.548 }, 00:09:38.548 { 00:09:38.548 "name": "BaseBdev3", 00:09:38.548 "uuid": "97c352a4-b8c3-41c3-bdaf-be42dc952efc", 00:09:38.548 "is_configured": true, 00:09:38.548 "data_offset": 2048, 00:09:38.548 "data_size": 63488 00:09:38.548 }, 00:09:38.548 { 00:09:38.548 "name": "BaseBdev4", 00:09:38.548 "uuid": "ca5447e8-2e17-43a5-88cf-54d20a67b7bf", 00:09:38.548 "is_configured": true, 00:09:38.548 "data_offset": 2048, 00:09:38.548 "data_size": 63488 00:09:38.548 } 00:09:38.548 ] 00:09:38.548 } 00:09:38.548 } 00:09:38.548 }' 00:09:38.548 16:47:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:38.548 BaseBdev2 00:09:38.548 BaseBdev3 00:09:38.548 BaseBdev4' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.548 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.808 [2024-09-29 16:48:00.237344] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:38.808 [2024-09-29 16:48:00.237419] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:38.808 [2024-09-29 16:48:00.237495] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:38.808 [2024-09-29 16:48:00.237558] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:38.808 [2024-09-29 16:48:00.237573] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 80690 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 80690 ']' 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 80690 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80690 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80690' 00:09:38.808 killing process with pid 80690 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 80690 00:09:38.808 [2024-09-29 16:48:00.271917] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:38.808 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 80690 00:09:38.808 [2024-09-29 16:48:00.311295] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:39.068 16:48:00 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:39.068 00:09:39.068 real 0m9.667s 00:09:39.068 user 0m16.672s 00:09:39.068 sys 0m1.884s 00:09:39.068 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:39.068 ************************************ 00:09:39.068 END TEST raid_state_function_test_sb 00:09:39.068 ************************************ 00:09:39.068 16:48:00 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.068 16:48:00 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:09:39.068 16:48:00 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:09:39.068 16:48:00 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:39.068 16:48:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:39.068 ************************************ 00:09:39.068 START TEST raid_superblock_test 00:09:39.068 ************************************ 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 4 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81348 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:39.068 16:48:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81348 00:09:39.069 16:48:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 81348 ']' 00:09:39.069 16:48:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:39.069 16:48:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:39.069 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:39.069 16:48:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:39.069 16:48:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:39.069 16:48:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.069 [2024-09-29 16:48:00.709907] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:39.069 [2024-09-29 16:48:00.710105] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81348 ] 00:09:39.328 [2024-09-29 16:48:00.854242] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:39.328 [2024-09-29 16:48:00.900898] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:39.328 [2024-09-29 16:48:00.942710] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:39.328 [2024-09-29 16:48:00.942832] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.897 malloc1 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:39.897 [2024-09-29 16:48:01.544747] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:39.897 [2024-09-29 16:48:01.544807] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:39.897 [2024-09-29 16:48:01.544832] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:39.897 [2024-09-29 16:48:01.544846] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:39.897 [2024-09-29 16:48:01.546921] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:39.897 [2024-09-29 16:48:01.546977] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:39.897 pt1 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.897 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.158 malloc2 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.158 [2024-09-29 16:48:01.594810] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:40.158 [2024-09-29 16:48:01.595017] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:40.158 [2024-09-29 16:48:01.595104] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:40.158 [2024-09-29 16:48:01.595210] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:40.158 [2024-09-29 16:48:01.599249] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:40.158 [2024-09-29 16:48:01.599357] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:40.158 pt2 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.158 malloc3 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.158 [2024-09-29 16:48:01.628288] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:40.158 [2024-09-29 16:48:01.628402] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:40.158 [2024-09-29 16:48:01.628437] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:40.158 [2024-09-29 16:48:01.628466] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:40.158 [2024-09-29 16:48:01.630489] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:40.158 [2024-09-29 16:48:01.630576] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:40.158 pt3 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.158 malloc4 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.158 [2024-09-29 16:48:01.660748] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:40.158 [2024-09-29 16:48:01.660792] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:40.158 [2024-09-29 16:48:01.660807] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:40.158 [2024-09-29 16:48:01.660819] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:40.158 [2024-09-29 16:48:01.662807] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:40.158 [2024-09-29 16:48:01.662889] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:40.158 pt4 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.158 [2024-09-29 16:48:01.672811] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:40.158 [2024-09-29 16:48:01.674635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:40.158 [2024-09-29 16:48:01.674700] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:40.158 [2024-09-29 16:48:01.674759] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:40.158 [2024-09-29 16:48:01.674902] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:40.158 [2024-09-29 16:48:01.674917] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:40.158 [2024-09-29 16:48:01.675184] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:40.158 [2024-09-29 16:48:01.675322] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:40.158 [2024-09-29 16:48:01.675332] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:40.158 [2024-09-29 16:48:01.675455] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.158 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.159 "name": "raid_bdev1", 00:09:40.159 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:40.159 "strip_size_kb": 64, 00:09:40.159 "state": "online", 00:09:40.159 "raid_level": "raid0", 00:09:40.159 "superblock": true, 00:09:40.159 "num_base_bdevs": 4, 00:09:40.159 "num_base_bdevs_discovered": 4, 00:09:40.159 "num_base_bdevs_operational": 4, 00:09:40.159 "base_bdevs_list": [ 00:09:40.159 { 00:09:40.159 "name": "pt1", 00:09:40.159 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:40.159 "is_configured": true, 00:09:40.159 "data_offset": 2048, 00:09:40.159 "data_size": 63488 00:09:40.159 }, 00:09:40.159 { 00:09:40.159 "name": "pt2", 00:09:40.159 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:40.159 "is_configured": true, 00:09:40.159 "data_offset": 2048, 00:09:40.159 "data_size": 63488 00:09:40.159 }, 00:09:40.159 { 00:09:40.159 "name": "pt3", 00:09:40.159 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:40.159 "is_configured": true, 00:09:40.159 "data_offset": 2048, 00:09:40.159 "data_size": 63488 00:09:40.159 }, 00:09:40.159 { 00:09:40.159 "name": "pt4", 00:09:40.159 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:40.159 "is_configured": true, 00:09:40.159 "data_offset": 2048, 00:09:40.159 "data_size": 63488 00:09:40.159 } 00:09:40.159 ] 00:09:40.159 }' 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.159 16:48:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.729 [2024-09-29 16:48:02.128260] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:40.729 "name": "raid_bdev1", 00:09:40.729 "aliases": [ 00:09:40.729 "af213fa3-ac92-492c-b52a-0b6b03dbc462" 00:09:40.729 ], 00:09:40.729 "product_name": "Raid Volume", 00:09:40.729 "block_size": 512, 00:09:40.729 "num_blocks": 253952, 00:09:40.729 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:40.729 "assigned_rate_limits": { 00:09:40.729 "rw_ios_per_sec": 0, 00:09:40.729 "rw_mbytes_per_sec": 0, 00:09:40.729 "r_mbytes_per_sec": 0, 00:09:40.729 "w_mbytes_per_sec": 0 00:09:40.729 }, 00:09:40.729 "claimed": false, 00:09:40.729 "zoned": false, 00:09:40.729 "supported_io_types": { 00:09:40.729 "read": true, 00:09:40.729 "write": true, 00:09:40.729 "unmap": true, 00:09:40.729 "flush": true, 00:09:40.729 "reset": true, 00:09:40.729 "nvme_admin": false, 00:09:40.729 "nvme_io": false, 00:09:40.729 "nvme_io_md": false, 00:09:40.729 "write_zeroes": true, 00:09:40.729 "zcopy": false, 00:09:40.729 "get_zone_info": false, 00:09:40.729 "zone_management": false, 00:09:40.729 "zone_append": false, 00:09:40.729 "compare": false, 00:09:40.729 "compare_and_write": false, 00:09:40.729 "abort": false, 00:09:40.729 "seek_hole": false, 00:09:40.729 "seek_data": false, 00:09:40.729 "copy": false, 00:09:40.729 "nvme_iov_md": false 00:09:40.729 }, 00:09:40.729 "memory_domains": [ 00:09:40.729 { 00:09:40.729 "dma_device_id": "system", 00:09:40.729 "dma_device_type": 1 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.729 "dma_device_type": 2 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "dma_device_id": "system", 00:09:40.729 "dma_device_type": 1 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.729 "dma_device_type": 2 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "dma_device_id": "system", 00:09:40.729 "dma_device_type": 1 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.729 "dma_device_type": 2 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "dma_device_id": "system", 00:09:40.729 "dma_device_type": 1 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:40.729 "dma_device_type": 2 00:09:40.729 } 00:09:40.729 ], 00:09:40.729 "driver_specific": { 00:09:40.729 "raid": { 00:09:40.729 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:40.729 "strip_size_kb": 64, 00:09:40.729 "state": "online", 00:09:40.729 "raid_level": "raid0", 00:09:40.729 "superblock": true, 00:09:40.729 "num_base_bdevs": 4, 00:09:40.729 "num_base_bdevs_discovered": 4, 00:09:40.729 "num_base_bdevs_operational": 4, 00:09:40.729 "base_bdevs_list": [ 00:09:40.729 { 00:09:40.729 "name": "pt1", 00:09:40.729 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:40.729 "is_configured": true, 00:09:40.729 "data_offset": 2048, 00:09:40.729 "data_size": 63488 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "name": "pt2", 00:09:40.729 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:40.729 "is_configured": true, 00:09:40.729 "data_offset": 2048, 00:09:40.729 "data_size": 63488 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "name": "pt3", 00:09:40.729 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:40.729 "is_configured": true, 00:09:40.729 "data_offset": 2048, 00:09:40.729 "data_size": 63488 00:09:40.729 }, 00:09:40.729 { 00:09:40.729 "name": "pt4", 00:09:40.729 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:40.729 "is_configured": true, 00:09:40.729 "data_offset": 2048, 00:09:40.729 "data_size": 63488 00:09:40.729 } 00:09:40.729 ] 00:09:40.729 } 00:09:40.729 } 00:09:40.729 }' 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:40.729 pt2 00:09:40.729 pt3 00:09:40.729 pt4' 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:40.729 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.730 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 [2024-09-29 16:48:02.459622] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=af213fa3-ac92-492c-b52a-0b6b03dbc462 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z af213fa3-ac92-492c-b52a-0b6b03dbc462 ']' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 [2024-09-29 16:48:02.491304] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:40.990 [2024-09-29 16:48:02.491370] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:40.990 [2024-09-29 16:48:02.491460] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:40.990 [2024-09-29 16:48:02.491528] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:40.990 [2024-09-29 16:48:02.491558] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:40.990 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:40.991 [2024-09-29 16:48:02.643087] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:40.991 [2024-09-29 16:48:02.644914] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:40.991 [2024-09-29 16:48:02.644960] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:40.991 [2024-09-29 16:48:02.644995] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:09:40.991 [2024-09-29 16:48:02.645043] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:40.991 [2024-09-29 16:48:02.645092] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:40.991 [2024-09-29 16:48:02.645113] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:40.991 [2024-09-29 16:48:02.645128] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:09:40.991 [2024-09-29 16:48:02.645142] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:40.991 [2024-09-29 16:48:02.645151] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:40.991 request: 00:09:40.991 { 00:09:40.991 "name": "raid_bdev1", 00:09:40.991 "raid_level": "raid0", 00:09:40.991 "base_bdevs": [ 00:09:40.991 "malloc1", 00:09:40.991 "malloc2", 00:09:40.991 "malloc3", 00:09:40.991 "malloc4" 00:09:40.991 ], 00:09:40.991 "strip_size_kb": 64, 00:09:40.991 "superblock": false, 00:09:40.991 "method": "bdev_raid_create", 00:09:40.991 "req_id": 1 00:09:40.991 } 00:09:40.991 Got JSON-RPC error response 00:09:40.991 response: 00:09:40.991 { 00:09:40.991 "code": -17, 00:09:40.991 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:40.991 } 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.991 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.250 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.250 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:41.250 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.251 [2024-09-29 16:48:02.710933] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:41.251 [2024-09-29 16:48:02.711022] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.251 [2024-09-29 16:48:02.711061] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:41.251 [2024-09-29 16:48:02.711088] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.251 [2024-09-29 16:48:02.713167] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.251 [2024-09-29 16:48:02.713233] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:41.251 [2024-09-29 16:48:02.713332] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:41.251 [2024-09-29 16:48:02.713389] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:41.251 pt1 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.251 "name": "raid_bdev1", 00:09:41.251 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:41.251 "strip_size_kb": 64, 00:09:41.251 "state": "configuring", 00:09:41.251 "raid_level": "raid0", 00:09:41.251 "superblock": true, 00:09:41.251 "num_base_bdevs": 4, 00:09:41.251 "num_base_bdevs_discovered": 1, 00:09:41.251 "num_base_bdevs_operational": 4, 00:09:41.251 "base_bdevs_list": [ 00:09:41.251 { 00:09:41.251 "name": "pt1", 00:09:41.251 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:41.251 "is_configured": true, 00:09:41.251 "data_offset": 2048, 00:09:41.251 "data_size": 63488 00:09:41.251 }, 00:09:41.251 { 00:09:41.251 "name": null, 00:09:41.251 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:41.251 "is_configured": false, 00:09:41.251 "data_offset": 2048, 00:09:41.251 "data_size": 63488 00:09:41.251 }, 00:09:41.251 { 00:09:41.251 "name": null, 00:09:41.251 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:41.251 "is_configured": false, 00:09:41.251 "data_offset": 2048, 00:09:41.251 "data_size": 63488 00:09:41.251 }, 00:09:41.251 { 00:09:41.251 "name": null, 00:09:41.251 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:41.251 "is_configured": false, 00:09:41.251 "data_offset": 2048, 00:09:41.251 "data_size": 63488 00:09:41.251 } 00:09:41.251 ] 00:09:41.251 }' 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.251 16:48:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.510 [2024-09-29 16:48:03.174168] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:41.510 [2024-09-29 16:48:03.174271] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:41.510 [2024-09-29 16:48:03.174297] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:09:41.510 [2024-09-29 16:48:03.174307] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:41.510 [2024-09-29 16:48:03.174680] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:41.510 [2024-09-29 16:48:03.174697] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:41.510 [2024-09-29 16:48:03.174785] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:41.510 [2024-09-29 16:48:03.174815] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:41.510 pt2 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.510 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.770 [2024-09-29 16:48:03.186175] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.770 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.770 "name": "raid_bdev1", 00:09:41.770 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:41.770 "strip_size_kb": 64, 00:09:41.770 "state": "configuring", 00:09:41.770 "raid_level": "raid0", 00:09:41.770 "superblock": true, 00:09:41.770 "num_base_bdevs": 4, 00:09:41.770 "num_base_bdevs_discovered": 1, 00:09:41.770 "num_base_bdevs_operational": 4, 00:09:41.770 "base_bdevs_list": [ 00:09:41.770 { 00:09:41.770 "name": "pt1", 00:09:41.770 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:41.770 "is_configured": true, 00:09:41.770 "data_offset": 2048, 00:09:41.770 "data_size": 63488 00:09:41.770 }, 00:09:41.770 { 00:09:41.770 "name": null, 00:09:41.770 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:41.770 "is_configured": false, 00:09:41.770 "data_offset": 0, 00:09:41.770 "data_size": 63488 00:09:41.770 }, 00:09:41.770 { 00:09:41.770 "name": null, 00:09:41.770 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:41.770 "is_configured": false, 00:09:41.770 "data_offset": 2048, 00:09:41.770 "data_size": 63488 00:09:41.770 }, 00:09:41.770 { 00:09:41.770 "name": null, 00:09:41.770 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:41.770 "is_configured": false, 00:09:41.771 "data_offset": 2048, 00:09:41.771 "data_size": 63488 00:09:41.771 } 00:09:41.771 ] 00:09:41.771 }' 00:09:41.771 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.771 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.030 [2024-09-29 16:48:03.605429] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:42.030 [2024-09-29 16:48:03.605530] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:42.030 [2024-09-29 16:48:03.605550] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:42.030 [2024-09-29 16:48:03.605560] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:42.030 [2024-09-29 16:48:03.605936] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:42.030 [2024-09-29 16:48:03.605955] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:42.030 [2024-09-29 16:48:03.606016] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:42.030 [2024-09-29 16:48:03.606038] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:42.030 pt2 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.030 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.030 [2024-09-29 16:48:03.617376] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:42.030 [2024-09-29 16:48:03.617427] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:42.030 [2024-09-29 16:48:03.617444] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:42.031 [2024-09-29 16:48:03.617461] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:42.031 [2024-09-29 16:48:03.617786] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:42.031 [2024-09-29 16:48:03.617804] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:42.031 [2024-09-29 16:48:03.617860] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:42.031 [2024-09-29 16:48:03.617880] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:42.031 pt3 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.031 [2024-09-29 16:48:03.629378] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:09:42.031 [2024-09-29 16:48:03.629423] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:42.031 [2024-09-29 16:48:03.629435] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:09:42.031 [2024-09-29 16:48:03.629444] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:42.031 [2024-09-29 16:48:03.629698] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:42.031 [2024-09-29 16:48:03.629715] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:09:42.031 [2024-09-29 16:48:03.629795] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:09:42.031 [2024-09-29 16:48:03.629814] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:09:42.031 [2024-09-29 16:48:03.629899] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:42.031 [2024-09-29 16:48:03.629909] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:42.031 [2024-09-29 16:48:03.630148] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:42.031 [2024-09-29 16:48:03.630274] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:42.031 [2024-09-29 16:48:03.630282] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:42.031 [2024-09-29 16:48:03.630374] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:42.031 pt4 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.031 "name": "raid_bdev1", 00:09:42.031 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:42.031 "strip_size_kb": 64, 00:09:42.031 "state": "online", 00:09:42.031 "raid_level": "raid0", 00:09:42.031 "superblock": true, 00:09:42.031 "num_base_bdevs": 4, 00:09:42.031 "num_base_bdevs_discovered": 4, 00:09:42.031 "num_base_bdevs_operational": 4, 00:09:42.031 "base_bdevs_list": [ 00:09:42.031 { 00:09:42.031 "name": "pt1", 00:09:42.031 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:42.031 "is_configured": true, 00:09:42.031 "data_offset": 2048, 00:09:42.031 "data_size": 63488 00:09:42.031 }, 00:09:42.031 { 00:09:42.031 "name": "pt2", 00:09:42.031 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:42.031 "is_configured": true, 00:09:42.031 "data_offset": 2048, 00:09:42.031 "data_size": 63488 00:09:42.031 }, 00:09:42.031 { 00:09:42.031 "name": "pt3", 00:09:42.031 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:42.031 "is_configured": true, 00:09:42.031 "data_offset": 2048, 00:09:42.031 "data_size": 63488 00:09:42.031 }, 00:09:42.031 { 00:09:42.031 "name": "pt4", 00:09:42.031 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:42.031 "is_configured": true, 00:09:42.031 "data_offset": 2048, 00:09:42.031 "data_size": 63488 00:09:42.031 } 00:09:42.031 ] 00:09:42.031 }' 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.031 16:48:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.602 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:42.602 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:42.602 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:42.602 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:42.602 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:42.602 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:42.602 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:42.603 [2024-09-29 16:48:04.116907] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:42.603 "name": "raid_bdev1", 00:09:42.603 "aliases": [ 00:09:42.603 "af213fa3-ac92-492c-b52a-0b6b03dbc462" 00:09:42.603 ], 00:09:42.603 "product_name": "Raid Volume", 00:09:42.603 "block_size": 512, 00:09:42.603 "num_blocks": 253952, 00:09:42.603 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:42.603 "assigned_rate_limits": { 00:09:42.603 "rw_ios_per_sec": 0, 00:09:42.603 "rw_mbytes_per_sec": 0, 00:09:42.603 "r_mbytes_per_sec": 0, 00:09:42.603 "w_mbytes_per_sec": 0 00:09:42.603 }, 00:09:42.603 "claimed": false, 00:09:42.603 "zoned": false, 00:09:42.603 "supported_io_types": { 00:09:42.603 "read": true, 00:09:42.603 "write": true, 00:09:42.603 "unmap": true, 00:09:42.603 "flush": true, 00:09:42.603 "reset": true, 00:09:42.603 "nvme_admin": false, 00:09:42.603 "nvme_io": false, 00:09:42.603 "nvme_io_md": false, 00:09:42.603 "write_zeroes": true, 00:09:42.603 "zcopy": false, 00:09:42.603 "get_zone_info": false, 00:09:42.603 "zone_management": false, 00:09:42.603 "zone_append": false, 00:09:42.603 "compare": false, 00:09:42.603 "compare_and_write": false, 00:09:42.603 "abort": false, 00:09:42.603 "seek_hole": false, 00:09:42.603 "seek_data": false, 00:09:42.603 "copy": false, 00:09:42.603 "nvme_iov_md": false 00:09:42.603 }, 00:09:42.603 "memory_domains": [ 00:09:42.603 { 00:09:42.603 "dma_device_id": "system", 00:09:42.603 "dma_device_type": 1 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.603 "dma_device_type": 2 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "dma_device_id": "system", 00:09:42.603 "dma_device_type": 1 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.603 "dma_device_type": 2 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "dma_device_id": "system", 00:09:42.603 "dma_device_type": 1 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.603 "dma_device_type": 2 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "dma_device_id": "system", 00:09:42.603 "dma_device_type": 1 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.603 "dma_device_type": 2 00:09:42.603 } 00:09:42.603 ], 00:09:42.603 "driver_specific": { 00:09:42.603 "raid": { 00:09:42.603 "uuid": "af213fa3-ac92-492c-b52a-0b6b03dbc462", 00:09:42.603 "strip_size_kb": 64, 00:09:42.603 "state": "online", 00:09:42.603 "raid_level": "raid0", 00:09:42.603 "superblock": true, 00:09:42.603 "num_base_bdevs": 4, 00:09:42.603 "num_base_bdevs_discovered": 4, 00:09:42.603 "num_base_bdevs_operational": 4, 00:09:42.603 "base_bdevs_list": [ 00:09:42.603 { 00:09:42.603 "name": "pt1", 00:09:42.603 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:42.603 "is_configured": true, 00:09:42.603 "data_offset": 2048, 00:09:42.603 "data_size": 63488 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "name": "pt2", 00:09:42.603 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:42.603 "is_configured": true, 00:09:42.603 "data_offset": 2048, 00:09:42.603 "data_size": 63488 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "name": "pt3", 00:09:42.603 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:42.603 "is_configured": true, 00:09:42.603 "data_offset": 2048, 00:09:42.603 "data_size": 63488 00:09:42.603 }, 00:09:42.603 { 00:09:42.603 "name": "pt4", 00:09:42.603 "uuid": "00000000-0000-0000-0000-000000000004", 00:09:42.603 "is_configured": true, 00:09:42.603 "data_offset": 2048, 00:09:42.603 "data_size": 63488 00:09:42.603 } 00:09:42.603 ] 00:09:42.603 } 00:09:42.603 } 00:09:42.603 }' 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:42.603 pt2 00:09:42.603 pt3 00:09:42.603 pt4' 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.603 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.863 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:42.864 [2024-09-29 16:48:04.416367] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' af213fa3-ac92-492c-b52a-0b6b03dbc462 '!=' af213fa3-ac92-492c-b52a-0b6b03dbc462 ']' 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81348 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 81348 ']' 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 81348 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81348 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81348' 00:09:42.864 killing process with pid 81348 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 81348 00:09:42.864 [2024-09-29 16:48:04.477066] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:42.864 [2024-09-29 16:48:04.477194] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:42.864 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 81348 00:09:42.864 [2024-09-29 16:48:04.477297] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:42.864 [2024-09-29 16:48:04.477312] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:42.864 [2024-09-29 16:48:04.519354] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:43.123 16:48:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:43.123 00:09:43.123 real 0m4.125s 00:09:43.123 user 0m6.498s 00:09:43.123 sys 0m0.877s 00:09:43.123 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:43.123 16:48:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.123 ************************************ 00:09:43.123 END TEST raid_superblock_test 00:09:43.123 ************************************ 00:09:43.384 16:48:04 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:09:43.384 16:48:04 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:43.384 16:48:04 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:43.384 16:48:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:43.384 ************************************ 00:09:43.384 START TEST raid_read_error_test 00:09:43.384 ************************************ 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 read 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.pPca81VvhV 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81592 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81592 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 81592 ']' 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:43.384 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:43.384 16:48:04 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.384 [2024-09-29 16:48:04.928178] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:43.384 [2024-09-29 16:48:04.928351] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81592 ] 00:09:43.644 [2024-09-29 16:48:05.072126] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:43.644 [2024-09-29 16:48:05.117979] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:43.644 [2024-09-29 16:48:05.159887] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:43.644 [2024-09-29 16:48:05.160000] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.214 BaseBdev1_malloc 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.214 true 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.214 [2024-09-29 16:48:05.778386] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:44.214 [2024-09-29 16:48:05.778500] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.214 [2024-09-29 16:48:05.778532] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:44.214 [2024-09-29 16:48:05.778547] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.214 [2024-09-29 16:48:05.780632] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.214 [2024-09-29 16:48:05.780669] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:44.214 BaseBdev1 00:09:44.214 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.215 BaseBdev2_malloc 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.215 true 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.215 [2024-09-29 16:48:05.835158] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:44.215 [2024-09-29 16:48:05.835226] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.215 [2024-09-29 16:48:05.835253] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:44.215 [2024-09-29 16:48:05.835264] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.215 [2024-09-29 16:48:05.838003] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.215 [2024-09-29 16:48:05.838047] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:44.215 BaseBdev2 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.215 BaseBdev3_malloc 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.215 true 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.215 [2024-09-29 16:48:05.875545] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:44.215 [2024-09-29 16:48:05.875645] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.215 [2024-09-29 16:48:05.875668] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:44.215 [2024-09-29 16:48:05.875676] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.215 [2024-09-29 16:48:05.877690] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.215 [2024-09-29 16:48:05.877744] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:44.215 BaseBdev3 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.215 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.474 BaseBdev4_malloc 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.474 true 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.474 [2024-09-29 16:48:05.916126] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:44.474 [2024-09-29 16:48:05.916227] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.474 [2024-09-29 16:48:05.916262] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:44.474 [2024-09-29 16:48:05.916271] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.474 [2024-09-29 16:48:05.918284] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.474 [2024-09-29 16:48:05.918319] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:44.474 BaseBdev4 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.474 [2024-09-29 16:48:05.928151] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:44.474 [2024-09-29 16:48:05.929968] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:44.474 [2024-09-29 16:48:05.930041] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:44.474 [2024-09-29 16:48:05.930101] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:44.474 [2024-09-29 16:48:05.930289] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:44.474 [2024-09-29 16:48:05.930300] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:44.474 [2024-09-29 16:48:05.930530] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:44.474 [2024-09-29 16:48:05.930651] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:44.474 [2024-09-29 16:48:05.930664] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:44.474 [2024-09-29 16:48:05.930798] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.474 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.475 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.475 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.475 "name": "raid_bdev1", 00:09:44.475 "uuid": "dfa72fee-9d2a-40eb-80ef-26ad6ffbaec2", 00:09:44.475 "strip_size_kb": 64, 00:09:44.475 "state": "online", 00:09:44.475 "raid_level": "raid0", 00:09:44.475 "superblock": true, 00:09:44.475 "num_base_bdevs": 4, 00:09:44.475 "num_base_bdevs_discovered": 4, 00:09:44.475 "num_base_bdevs_operational": 4, 00:09:44.475 "base_bdevs_list": [ 00:09:44.475 { 00:09:44.475 "name": "BaseBdev1", 00:09:44.475 "uuid": "21f315e0-1ad3-569f-adb5-fbf33b445484", 00:09:44.475 "is_configured": true, 00:09:44.475 "data_offset": 2048, 00:09:44.475 "data_size": 63488 00:09:44.475 }, 00:09:44.475 { 00:09:44.475 "name": "BaseBdev2", 00:09:44.475 "uuid": "4442168f-9b1b-5362-a4b0-a057d4beb401", 00:09:44.475 "is_configured": true, 00:09:44.475 "data_offset": 2048, 00:09:44.475 "data_size": 63488 00:09:44.475 }, 00:09:44.475 { 00:09:44.475 "name": "BaseBdev3", 00:09:44.475 "uuid": "268e0019-cf6b-5d8e-9830-5c501f204284", 00:09:44.475 "is_configured": true, 00:09:44.475 "data_offset": 2048, 00:09:44.475 "data_size": 63488 00:09:44.475 }, 00:09:44.475 { 00:09:44.475 "name": "BaseBdev4", 00:09:44.475 "uuid": "aa8e7fe8-3ee9-51cd-942d-c3e3f73615cd", 00:09:44.475 "is_configured": true, 00:09:44.475 "data_offset": 2048, 00:09:44.475 "data_size": 63488 00:09:44.475 } 00:09:44.475 ] 00:09:44.475 }' 00:09:44.475 16:48:05 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.475 16:48:05 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.735 16:48:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:44.735 16:48:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:44.996 [2024-09-29 16:48:06.419690] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:45.936 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.937 "name": "raid_bdev1", 00:09:45.937 "uuid": "dfa72fee-9d2a-40eb-80ef-26ad6ffbaec2", 00:09:45.937 "strip_size_kb": 64, 00:09:45.937 "state": "online", 00:09:45.937 "raid_level": "raid0", 00:09:45.937 "superblock": true, 00:09:45.937 "num_base_bdevs": 4, 00:09:45.937 "num_base_bdevs_discovered": 4, 00:09:45.937 "num_base_bdevs_operational": 4, 00:09:45.937 "base_bdevs_list": [ 00:09:45.937 { 00:09:45.937 "name": "BaseBdev1", 00:09:45.937 "uuid": "21f315e0-1ad3-569f-adb5-fbf33b445484", 00:09:45.937 "is_configured": true, 00:09:45.937 "data_offset": 2048, 00:09:45.937 "data_size": 63488 00:09:45.937 }, 00:09:45.937 { 00:09:45.937 "name": "BaseBdev2", 00:09:45.937 "uuid": "4442168f-9b1b-5362-a4b0-a057d4beb401", 00:09:45.937 "is_configured": true, 00:09:45.937 "data_offset": 2048, 00:09:45.937 "data_size": 63488 00:09:45.937 }, 00:09:45.937 { 00:09:45.937 "name": "BaseBdev3", 00:09:45.937 "uuid": "268e0019-cf6b-5d8e-9830-5c501f204284", 00:09:45.937 "is_configured": true, 00:09:45.937 "data_offset": 2048, 00:09:45.937 "data_size": 63488 00:09:45.937 }, 00:09:45.937 { 00:09:45.937 "name": "BaseBdev4", 00:09:45.937 "uuid": "aa8e7fe8-3ee9-51cd-942d-c3e3f73615cd", 00:09:45.937 "is_configured": true, 00:09:45.937 "data_offset": 2048, 00:09:45.937 "data_size": 63488 00:09:45.937 } 00:09:45.937 ] 00:09:45.937 }' 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.937 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.197 [2024-09-29 16:48:07.815289] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:46.197 [2024-09-29 16:48:07.815380] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:46.197 [2024-09-29 16:48:07.817966] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:46.197 [2024-09-29 16:48:07.818054] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:46.197 [2024-09-29 16:48:07.818116] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:46.197 [2024-09-29 16:48:07.818172] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:46.197 { 00:09:46.197 "results": [ 00:09:46.197 { 00:09:46.197 "job": "raid_bdev1", 00:09:46.197 "core_mask": "0x1", 00:09:46.197 "workload": "randrw", 00:09:46.197 "percentage": 50, 00:09:46.197 "status": "finished", 00:09:46.197 "queue_depth": 1, 00:09:46.197 "io_size": 131072, 00:09:46.197 "runtime": 1.396476, 00:09:46.197 "iops": 17087.29688157906, 00:09:46.197 "mibps": 2135.9121101973824, 00:09:46.197 "io_failed": 1, 00:09:46.197 "io_timeout": 0, 00:09:46.197 "avg_latency_us": 81.17869285497437, 00:09:46.197 "min_latency_us": 24.593886462882097, 00:09:46.197 "max_latency_us": 1352.216593886463 00:09:46.197 } 00:09:46.197 ], 00:09:46.197 "core_count": 1 00:09:46.197 } 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81592 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 81592 ']' 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 81592 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81592 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81592' 00:09:46.197 killing process with pid 81592 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 81592 00:09:46.197 [2024-09-29 16:48:07.865068] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:46.197 16:48:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 81592 00:09:46.458 [2024-09-29 16:48:07.899169] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:46.458 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.pPca81VvhV 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:09:46.718 00:09:46.718 real 0m3.316s 00:09:46.718 user 0m4.159s 00:09:46.718 sys 0m0.534s 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:46.718 ************************************ 00:09:46.718 END TEST raid_read_error_test 00:09:46.718 ************************************ 00:09:46.718 16:48:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.718 16:48:08 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:09:46.718 16:48:08 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:46.718 16:48:08 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:46.718 16:48:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:46.718 ************************************ 00:09:46.718 START TEST raid_write_error_test 00:09:46.718 ************************************ 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 write 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.OZfqSwc1nu 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81726 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81726 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 81726 ']' 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:46.718 16:48:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:46.719 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:46.719 16:48:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:46.719 16:48:08 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.719 [2024-09-29 16:48:08.309639] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:46.719 [2024-09-29 16:48:08.309782] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81726 ] 00:09:46.978 [2024-09-29 16:48:08.455554] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:46.978 [2024-09-29 16:48:08.503390] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:46.978 [2024-09-29 16:48:08.545280] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:46.978 [2024-09-29 16:48:08.545313] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:47.553 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.554 BaseBdev1_malloc 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.554 true 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.554 [2024-09-29 16:48:09.163022] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:47.554 [2024-09-29 16:48:09.163077] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.554 [2024-09-29 16:48:09.163101] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:47.554 [2024-09-29 16:48:09.163110] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.554 [2024-09-29 16:48:09.165223] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.554 [2024-09-29 16:48:09.165260] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:47.554 BaseBdev1 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.554 BaseBdev2_malloc 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.554 true 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.554 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.554 [2024-09-29 16:48:09.220033] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:47.554 [2024-09-29 16:48:09.220100] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.554 [2024-09-29 16:48:09.220126] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:47.554 [2024-09-29 16:48:09.220137] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.554 [2024-09-29 16:48:09.222750] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.554 [2024-09-29 16:48:09.222790] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:47.816 BaseBdev2 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.816 BaseBdev3_malloc 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.816 true 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.816 [2024-09-29 16:48:09.260626] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:47.816 [2024-09-29 16:48:09.260673] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.816 [2024-09-29 16:48:09.260691] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:47.816 [2024-09-29 16:48:09.260700] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.816 [2024-09-29 16:48:09.262750] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.816 [2024-09-29 16:48:09.262785] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:47.816 BaseBdev3 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.816 BaseBdev4_malloc 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.816 true 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.816 [2024-09-29 16:48:09.301243] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:09:47.816 [2024-09-29 16:48:09.301326] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.816 [2024-09-29 16:48:09.301364] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:47.816 [2024-09-29 16:48:09.301391] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.816 [2024-09-29 16:48:09.303426] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.816 [2024-09-29 16:48:09.303497] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:09:47.816 BaseBdev4 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.816 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.816 [2024-09-29 16:48:09.313278] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:47.816 [2024-09-29 16:48:09.315068] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:47.816 [2024-09-29 16:48:09.315142] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:47.816 [2024-09-29 16:48:09.315204] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:47.816 [2024-09-29 16:48:09.315385] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:09:47.816 [2024-09-29 16:48:09.315401] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:09:47.816 [2024-09-29 16:48:09.315647] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:47.817 [2024-09-29 16:48:09.315803] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:09:47.817 [2024-09-29 16:48:09.315816] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:09:47.817 [2024-09-29 16:48:09.315933] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.817 "name": "raid_bdev1", 00:09:47.817 "uuid": "ccdb9b06-b99a-45f6-96f9-4977a1afb1b5", 00:09:47.817 "strip_size_kb": 64, 00:09:47.817 "state": "online", 00:09:47.817 "raid_level": "raid0", 00:09:47.817 "superblock": true, 00:09:47.817 "num_base_bdevs": 4, 00:09:47.817 "num_base_bdevs_discovered": 4, 00:09:47.817 "num_base_bdevs_operational": 4, 00:09:47.817 "base_bdevs_list": [ 00:09:47.817 { 00:09:47.817 "name": "BaseBdev1", 00:09:47.817 "uuid": "e26935d9-9532-58b1-a2fc-c505a26fda97", 00:09:47.817 "is_configured": true, 00:09:47.817 "data_offset": 2048, 00:09:47.817 "data_size": 63488 00:09:47.817 }, 00:09:47.817 { 00:09:47.817 "name": "BaseBdev2", 00:09:47.817 "uuid": "150a47c7-b221-54dc-a578-eafa54eec0ff", 00:09:47.817 "is_configured": true, 00:09:47.817 "data_offset": 2048, 00:09:47.817 "data_size": 63488 00:09:47.817 }, 00:09:47.817 { 00:09:47.817 "name": "BaseBdev3", 00:09:47.817 "uuid": "4746a8bf-12d6-5ab6-aed2-ea99781cd9b2", 00:09:47.817 "is_configured": true, 00:09:47.817 "data_offset": 2048, 00:09:47.817 "data_size": 63488 00:09:47.817 }, 00:09:47.817 { 00:09:47.817 "name": "BaseBdev4", 00:09:47.817 "uuid": "92829b53-324a-5c11-a710-f4729229318e", 00:09:47.817 "is_configured": true, 00:09:47.817 "data_offset": 2048, 00:09:47.817 "data_size": 63488 00:09:47.817 } 00:09:47.817 ] 00:09:47.817 }' 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.817 16:48:09 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.076 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:48.076 16:48:09 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:48.336 [2024-09-29 16:48:09.832713] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:09:49.299 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:49.299 16:48:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.299 16:48:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.299 16:48:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.299 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:49.299 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:09:49.299 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.300 "name": "raid_bdev1", 00:09:49.300 "uuid": "ccdb9b06-b99a-45f6-96f9-4977a1afb1b5", 00:09:49.300 "strip_size_kb": 64, 00:09:49.300 "state": "online", 00:09:49.300 "raid_level": "raid0", 00:09:49.300 "superblock": true, 00:09:49.300 "num_base_bdevs": 4, 00:09:49.300 "num_base_bdevs_discovered": 4, 00:09:49.300 "num_base_bdevs_operational": 4, 00:09:49.300 "base_bdevs_list": [ 00:09:49.300 { 00:09:49.300 "name": "BaseBdev1", 00:09:49.300 "uuid": "e26935d9-9532-58b1-a2fc-c505a26fda97", 00:09:49.300 "is_configured": true, 00:09:49.300 "data_offset": 2048, 00:09:49.300 "data_size": 63488 00:09:49.300 }, 00:09:49.300 { 00:09:49.300 "name": "BaseBdev2", 00:09:49.300 "uuid": "150a47c7-b221-54dc-a578-eafa54eec0ff", 00:09:49.300 "is_configured": true, 00:09:49.300 "data_offset": 2048, 00:09:49.300 "data_size": 63488 00:09:49.300 }, 00:09:49.300 { 00:09:49.300 "name": "BaseBdev3", 00:09:49.300 "uuid": "4746a8bf-12d6-5ab6-aed2-ea99781cd9b2", 00:09:49.300 "is_configured": true, 00:09:49.300 "data_offset": 2048, 00:09:49.300 "data_size": 63488 00:09:49.300 }, 00:09:49.300 { 00:09:49.300 "name": "BaseBdev4", 00:09:49.300 "uuid": "92829b53-324a-5c11-a710-f4729229318e", 00:09:49.300 "is_configured": true, 00:09:49.300 "data_offset": 2048, 00:09:49.300 "data_size": 63488 00:09:49.300 } 00:09:49.300 ] 00:09:49.300 }' 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.300 16:48:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.569 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:49.569 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.569 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.569 [2024-09-29 16:48:11.236775] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:49.569 [2024-09-29 16:48:11.236807] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:49.569 [2024-09-29 16:48:11.239328] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:49.569 [2024-09-29 16:48:11.239440] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:49.569 [2024-09-29 16:48:11.239497] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:49.569 [2024-09-29 16:48:11.239507] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:09:49.569 { 00:09:49.569 "results": [ 00:09:49.569 { 00:09:49.570 "job": "raid_bdev1", 00:09:49.570 "core_mask": "0x1", 00:09:49.570 "workload": "randrw", 00:09:49.570 "percentage": 50, 00:09:49.570 "status": "finished", 00:09:49.570 "queue_depth": 1, 00:09:49.570 "io_size": 131072, 00:09:49.570 "runtime": 1.404914, 00:09:49.570 "iops": 16934.132623064474, 00:09:49.570 "mibps": 2116.766577883059, 00:09:49.570 "io_failed": 1, 00:09:49.570 "io_timeout": 0, 00:09:49.570 "avg_latency_us": 81.94360792075719, 00:09:49.570 "min_latency_us": 25.041048034934498, 00:09:49.570 "max_latency_us": 1373.6803493449781 00:09:49.570 } 00:09:49.570 ], 00:09:49.570 "core_count": 1 00:09:49.570 } 00:09:49.570 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.570 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81726 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 81726 ']' 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 81726 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81726 00:09:49.830 killing process with pid 81726 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81726' 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 81726 00:09:49.830 [2024-09-29 16:48:11.285095] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:49.830 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 81726 00:09:49.830 [2024-09-29 16:48:11.319210] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.OZfqSwc1nu 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.71 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:50.090 ************************************ 00:09:50.090 END TEST raid_write_error_test 00:09:50.090 ************************************ 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.71 != \0\.\0\0 ]] 00:09:50.090 00:09:50.090 real 0m3.350s 00:09:50.090 user 0m4.245s 00:09:50.090 sys 0m0.513s 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:50.090 16:48:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.090 16:48:11 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:50.090 16:48:11 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:09:50.090 16:48:11 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:50.090 16:48:11 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:50.090 16:48:11 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:50.090 ************************************ 00:09:50.090 START TEST raid_state_function_test 00:09:50.090 ************************************ 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 false 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=81859 00:09:50.090 Process raid pid: 81859 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 81859' 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 81859 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 81859 ']' 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:50.090 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:50.090 16:48:11 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.090 [2024-09-29 16:48:11.723057] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:50.090 [2024-09-29 16:48:11.723175] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:50.350 [2024-09-29 16:48:11.866984] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:50.350 [2024-09-29 16:48:11.912547] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:50.350 [2024-09-29 16:48:11.954643] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:50.350 [2024-09-29 16:48:11.954688] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.920 [2024-09-29 16:48:12.551960] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:50.920 [2024-09-29 16:48:12.552071] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:50.920 [2024-09-29 16:48:12.552097] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:50.920 [2024-09-29 16:48:12.552120] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:50.920 [2024-09-29 16:48:12.552126] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:50.920 [2024-09-29 16:48:12.552139] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:50.920 [2024-09-29 16:48:12.552144] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:50.920 [2024-09-29 16:48:12.552153] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.920 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.180 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.180 "name": "Existed_Raid", 00:09:51.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.180 "strip_size_kb": 64, 00:09:51.180 "state": "configuring", 00:09:51.180 "raid_level": "concat", 00:09:51.180 "superblock": false, 00:09:51.180 "num_base_bdevs": 4, 00:09:51.180 "num_base_bdevs_discovered": 0, 00:09:51.180 "num_base_bdevs_operational": 4, 00:09:51.180 "base_bdevs_list": [ 00:09:51.180 { 00:09:51.180 "name": "BaseBdev1", 00:09:51.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.180 "is_configured": false, 00:09:51.180 "data_offset": 0, 00:09:51.180 "data_size": 0 00:09:51.180 }, 00:09:51.180 { 00:09:51.180 "name": "BaseBdev2", 00:09:51.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.180 "is_configured": false, 00:09:51.180 "data_offset": 0, 00:09:51.180 "data_size": 0 00:09:51.180 }, 00:09:51.180 { 00:09:51.180 "name": "BaseBdev3", 00:09:51.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.180 "is_configured": false, 00:09:51.180 "data_offset": 0, 00:09:51.180 "data_size": 0 00:09:51.180 }, 00:09:51.180 { 00:09:51.180 "name": "BaseBdev4", 00:09:51.180 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.180 "is_configured": false, 00:09:51.180 "data_offset": 0, 00:09:51.180 "data_size": 0 00:09:51.180 } 00:09:51.180 ] 00:09:51.180 }' 00:09:51.180 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.180 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.440 [2024-09-29 16:48:12.979071] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:51.440 [2024-09-29 16:48:12.979150] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.440 [2024-09-29 16:48:12.991081] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:51.440 [2024-09-29 16:48:12.991158] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:51.440 [2024-09-29 16:48:12.991184] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:51.440 [2024-09-29 16:48:12.991206] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:51.440 [2024-09-29 16:48:12.991224] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:51.440 [2024-09-29 16:48:12.991244] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:51.440 [2024-09-29 16:48:12.991261] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:51.440 [2024-09-29 16:48:12.991281] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.440 16:48:12 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.440 [2024-09-29 16:48:13.011978] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:51.440 BaseBdev1 00:09:51.440 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.440 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:51.440 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.441 [ 00:09:51.441 { 00:09:51.441 "name": "BaseBdev1", 00:09:51.441 "aliases": [ 00:09:51.441 "d020532a-baa2-4e5a-91cb-c82ac7c6b484" 00:09:51.441 ], 00:09:51.441 "product_name": "Malloc disk", 00:09:51.441 "block_size": 512, 00:09:51.441 "num_blocks": 65536, 00:09:51.441 "uuid": "d020532a-baa2-4e5a-91cb-c82ac7c6b484", 00:09:51.441 "assigned_rate_limits": { 00:09:51.441 "rw_ios_per_sec": 0, 00:09:51.441 "rw_mbytes_per_sec": 0, 00:09:51.441 "r_mbytes_per_sec": 0, 00:09:51.441 "w_mbytes_per_sec": 0 00:09:51.441 }, 00:09:51.441 "claimed": true, 00:09:51.441 "claim_type": "exclusive_write", 00:09:51.441 "zoned": false, 00:09:51.441 "supported_io_types": { 00:09:51.441 "read": true, 00:09:51.441 "write": true, 00:09:51.441 "unmap": true, 00:09:51.441 "flush": true, 00:09:51.441 "reset": true, 00:09:51.441 "nvme_admin": false, 00:09:51.441 "nvme_io": false, 00:09:51.441 "nvme_io_md": false, 00:09:51.441 "write_zeroes": true, 00:09:51.441 "zcopy": true, 00:09:51.441 "get_zone_info": false, 00:09:51.441 "zone_management": false, 00:09:51.441 "zone_append": false, 00:09:51.441 "compare": false, 00:09:51.441 "compare_and_write": false, 00:09:51.441 "abort": true, 00:09:51.441 "seek_hole": false, 00:09:51.441 "seek_data": false, 00:09:51.441 "copy": true, 00:09:51.441 "nvme_iov_md": false 00:09:51.441 }, 00:09:51.441 "memory_domains": [ 00:09:51.441 { 00:09:51.441 "dma_device_id": "system", 00:09:51.441 "dma_device_type": 1 00:09:51.441 }, 00:09:51.441 { 00:09:51.441 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:51.441 "dma_device_type": 2 00:09:51.441 } 00:09:51.441 ], 00:09:51.441 "driver_specific": {} 00:09:51.441 } 00:09:51.441 ] 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.441 "name": "Existed_Raid", 00:09:51.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.441 "strip_size_kb": 64, 00:09:51.441 "state": "configuring", 00:09:51.441 "raid_level": "concat", 00:09:51.441 "superblock": false, 00:09:51.441 "num_base_bdevs": 4, 00:09:51.441 "num_base_bdevs_discovered": 1, 00:09:51.441 "num_base_bdevs_operational": 4, 00:09:51.441 "base_bdevs_list": [ 00:09:51.441 { 00:09:51.441 "name": "BaseBdev1", 00:09:51.441 "uuid": "d020532a-baa2-4e5a-91cb-c82ac7c6b484", 00:09:51.441 "is_configured": true, 00:09:51.441 "data_offset": 0, 00:09:51.441 "data_size": 65536 00:09:51.441 }, 00:09:51.441 { 00:09:51.441 "name": "BaseBdev2", 00:09:51.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.441 "is_configured": false, 00:09:51.441 "data_offset": 0, 00:09:51.441 "data_size": 0 00:09:51.441 }, 00:09:51.441 { 00:09:51.441 "name": "BaseBdev3", 00:09:51.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.441 "is_configured": false, 00:09:51.441 "data_offset": 0, 00:09:51.441 "data_size": 0 00:09:51.441 }, 00:09:51.441 { 00:09:51.441 "name": "BaseBdev4", 00:09:51.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:51.441 "is_configured": false, 00:09:51.441 "data_offset": 0, 00:09:51.441 "data_size": 0 00:09:51.441 } 00:09:51.441 ] 00:09:51.441 }' 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.441 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 [2024-09-29 16:48:13.471194] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:52.012 [2024-09-29 16:48:13.471236] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 [2024-09-29 16:48:13.483237] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:52.012 [2024-09-29 16:48:13.485066] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:52.012 [2024-09-29 16:48:13.485107] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:52.012 [2024-09-29 16:48:13.485115] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:52.012 [2024-09-29 16:48:13.485124] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:52.012 [2024-09-29 16:48:13.485130] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:52.012 [2024-09-29 16:48:13.485138] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.012 "name": "Existed_Raid", 00:09:52.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.012 "strip_size_kb": 64, 00:09:52.012 "state": "configuring", 00:09:52.012 "raid_level": "concat", 00:09:52.012 "superblock": false, 00:09:52.012 "num_base_bdevs": 4, 00:09:52.012 "num_base_bdevs_discovered": 1, 00:09:52.012 "num_base_bdevs_operational": 4, 00:09:52.012 "base_bdevs_list": [ 00:09:52.012 { 00:09:52.012 "name": "BaseBdev1", 00:09:52.012 "uuid": "d020532a-baa2-4e5a-91cb-c82ac7c6b484", 00:09:52.012 "is_configured": true, 00:09:52.012 "data_offset": 0, 00:09:52.012 "data_size": 65536 00:09:52.012 }, 00:09:52.012 { 00:09:52.012 "name": "BaseBdev2", 00:09:52.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.012 "is_configured": false, 00:09:52.012 "data_offset": 0, 00:09:52.012 "data_size": 0 00:09:52.012 }, 00:09:52.012 { 00:09:52.012 "name": "BaseBdev3", 00:09:52.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.012 "is_configured": false, 00:09:52.012 "data_offset": 0, 00:09:52.012 "data_size": 0 00:09:52.012 }, 00:09:52.012 { 00:09:52.012 "name": "BaseBdev4", 00:09:52.012 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.012 "is_configured": false, 00:09:52.012 "data_offset": 0, 00:09:52.012 "data_size": 0 00:09:52.012 } 00:09:52.012 ] 00:09:52.012 }' 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.012 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.273 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:52.273 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.273 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.534 [2024-09-29 16:48:13.952451] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:52.534 BaseBdev2 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.534 [ 00:09:52.534 { 00:09:52.534 "name": "BaseBdev2", 00:09:52.534 "aliases": [ 00:09:52.534 "734315f4-8aff-426c-b59d-65ef43126edb" 00:09:52.534 ], 00:09:52.534 "product_name": "Malloc disk", 00:09:52.534 "block_size": 512, 00:09:52.534 "num_blocks": 65536, 00:09:52.534 "uuid": "734315f4-8aff-426c-b59d-65ef43126edb", 00:09:52.534 "assigned_rate_limits": { 00:09:52.534 "rw_ios_per_sec": 0, 00:09:52.534 "rw_mbytes_per_sec": 0, 00:09:52.534 "r_mbytes_per_sec": 0, 00:09:52.534 "w_mbytes_per_sec": 0 00:09:52.534 }, 00:09:52.534 "claimed": true, 00:09:52.534 "claim_type": "exclusive_write", 00:09:52.534 "zoned": false, 00:09:52.534 "supported_io_types": { 00:09:52.534 "read": true, 00:09:52.534 "write": true, 00:09:52.534 "unmap": true, 00:09:52.534 "flush": true, 00:09:52.534 "reset": true, 00:09:52.534 "nvme_admin": false, 00:09:52.534 "nvme_io": false, 00:09:52.534 "nvme_io_md": false, 00:09:52.534 "write_zeroes": true, 00:09:52.534 "zcopy": true, 00:09:52.534 "get_zone_info": false, 00:09:52.534 "zone_management": false, 00:09:52.534 "zone_append": false, 00:09:52.534 "compare": false, 00:09:52.534 "compare_and_write": false, 00:09:52.534 "abort": true, 00:09:52.534 "seek_hole": false, 00:09:52.534 "seek_data": false, 00:09:52.534 "copy": true, 00:09:52.534 "nvme_iov_md": false 00:09:52.534 }, 00:09:52.534 "memory_domains": [ 00:09:52.534 { 00:09:52.534 "dma_device_id": "system", 00:09:52.534 "dma_device_type": 1 00:09:52.534 }, 00:09:52.534 { 00:09:52.534 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.534 "dma_device_type": 2 00:09:52.534 } 00:09:52.534 ], 00:09:52.534 "driver_specific": {} 00:09:52.534 } 00:09:52.534 ] 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.534 16:48:13 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.534 "name": "Existed_Raid", 00:09:52.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.534 "strip_size_kb": 64, 00:09:52.534 "state": "configuring", 00:09:52.534 "raid_level": "concat", 00:09:52.534 "superblock": false, 00:09:52.534 "num_base_bdevs": 4, 00:09:52.534 "num_base_bdevs_discovered": 2, 00:09:52.534 "num_base_bdevs_operational": 4, 00:09:52.534 "base_bdevs_list": [ 00:09:52.534 { 00:09:52.534 "name": "BaseBdev1", 00:09:52.534 "uuid": "d020532a-baa2-4e5a-91cb-c82ac7c6b484", 00:09:52.534 "is_configured": true, 00:09:52.534 "data_offset": 0, 00:09:52.534 "data_size": 65536 00:09:52.534 }, 00:09:52.534 { 00:09:52.534 "name": "BaseBdev2", 00:09:52.534 "uuid": "734315f4-8aff-426c-b59d-65ef43126edb", 00:09:52.534 "is_configured": true, 00:09:52.534 "data_offset": 0, 00:09:52.534 "data_size": 65536 00:09:52.534 }, 00:09:52.534 { 00:09:52.534 "name": "BaseBdev3", 00:09:52.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.534 "is_configured": false, 00:09:52.534 "data_offset": 0, 00:09:52.534 "data_size": 0 00:09:52.534 }, 00:09:52.534 { 00:09:52.534 "name": "BaseBdev4", 00:09:52.534 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:52.534 "is_configured": false, 00:09:52.534 "data_offset": 0, 00:09:52.534 "data_size": 0 00:09:52.534 } 00:09:52.534 ] 00:09:52.534 }' 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.534 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.795 [2024-09-29 16:48:14.398573] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:52.795 BaseBdev3 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.795 [ 00:09:52.795 { 00:09:52.795 "name": "BaseBdev3", 00:09:52.795 "aliases": [ 00:09:52.795 "d6f56f82-d766-4364-ab4f-54f0326bcb42" 00:09:52.795 ], 00:09:52.795 "product_name": "Malloc disk", 00:09:52.795 "block_size": 512, 00:09:52.795 "num_blocks": 65536, 00:09:52.795 "uuid": "d6f56f82-d766-4364-ab4f-54f0326bcb42", 00:09:52.795 "assigned_rate_limits": { 00:09:52.795 "rw_ios_per_sec": 0, 00:09:52.795 "rw_mbytes_per_sec": 0, 00:09:52.795 "r_mbytes_per_sec": 0, 00:09:52.795 "w_mbytes_per_sec": 0 00:09:52.795 }, 00:09:52.795 "claimed": true, 00:09:52.795 "claim_type": "exclusive_write", 00:09:52.795 "zoned": false, 00:09:52.795 "supported_io_types": { 00:09:52.795 "read": true, 00:09:52.795 "write": true, 00:09:52.795 "unmap": true, 00:09:52.795 "flush": true, 00:09:52.795 "reset": true, 00:09:52.795 "nvme_admin": false, 00:09:52.795 "nvme_io": false, 00:09:52.795 "nvme_io_md": false, 00:09:52.795 "write_zeroes": true, 00:09:52.795 "zcopy": true, 00:09:52.795 "get_zone_info": false, 00:09:52.795 "zone_management": false, 00:09:52.795 "zone_append": false, 00:09:52.795 "compare": false, 00:09:52.795 "compare_and_write": false, 00:09:52.795 "abort": true, 00:09:52.795 "seek_hole": false, 00:09:52.795 "seek_data": false, 00:09:52.795 "copy": true, 00:09:52.795 "nvme_iov_md": false 00:09:52.795 }, 00:09:52.795 "memory_domains": [ 00:09:52.795 { 00:09:52.795 "dma_device_id": "system", 00:09:52.795 "dma_device_type": 1 00:09:52.795 }, 00:09:52.795 { 00:09:52.795 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:52.795 "dma_device_type": 2 00:09:52.795 } 00:09:52.795 ], 00:09:52.795 "driver_specific": {} 00:09:52.795 } 00:09:52.795 ] 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.795 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.055 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.055 "name": "Existed_Raid", 00:09:53.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.055 "strip_size_kb": 64, 00:09:53.055 "state": "configuring", 00:09:53.055 "raid_level": "concat", 00:09:53.055 "superblock": false, 00:09:53.055 "num_base_bdevs": 4, 00:09:53.055 "num_base_bdevs_discovered": 3, 00:09:53.055 "num_base_bdevs_operational": 4, 00:09:53.055 "base_bdevs_list": [ 00:09:53.055 { 00:09:53.055 "name": "BaseBdev1", 00:09:53.055 "uuid": "d020532a-baa2-4e5a-91cb-c82ac7c6b484", 00:09:53.055 "is_configured": true, 00:09:53.055 "data_offset": 0, 00:09:53.055 "data_size": 65536 00:09:53.055 }, 00:09:53.055 { 00:09:53.055 "name": "BaseBdev2", 00:09:53.055 "uuid": "734315f4-8aff-426c-b59d-65ef43126edb", 00:09:53.055 "is_configured": true, 00:09:53.055 "data_offset": 0, 00:09:53.055 "data_size": 65536 00:09:53.055 }, 00:09:53.055 { 00:09:53.056 "name": "BaseBdev3", 00:09:53.056 "uuid": "d6f56f82-d766-4364-ab4f-54f0326bcb42", 00:09:53.056 "is_configured": true, 00:09:53.056 "data_offset": 0, 00:09:53.056 "data_size": 65536 00:09:53.056 }, 00:09:53.056 { 00:09:53.056 "name": "BaseBdev4", 00:09:53.056 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:53.056 "is_configured": false, 00:09:53.056 "data_offset": 0, 00:09:53.056 "data_size": 0 00:09:53.056 } 00:09:53.056 ] 00:09:53.056 }' 00:09:53.056 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.056 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.315 [2024-09-29 16:48:14.900716] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:53.315 [2024-09-29 16:48:14.900857] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:53.315 [2024-09-29 16:48:14.900882] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:53.315 [2024-09-29 16:48:14.901199] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:53.315 BaseBdev4 00:09:53.315 [2024-09-29 16:48:14.901394] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:53.315 [2024-09-29 16:48:14.901419] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:53.315 [2024-09-29 16:48:14.901618] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.315 [ 00:09:53.315 { 00:09:53.315 "name": "BaseBdev4", 00:09:53.315 "aliases": [ 00:09:53.315 "cd937bf8-0bb5-4626-bd77-415c929d75c9" 00:09:53.315 ], 00:09:53.315 "product_name": "Malloc disk", 00:09:53.315 "block_size": 512, 00:09:53.315 "num_blocks": 65536, 00:09:53.315 "uuid": "cd937bf8-0bb5-4626-bd77-415c929d75c9", 00:09:53.315 "assigned_rate_limits": { 00:09:53.315 "rw_ios_per_sec": 0, 00:09:53.315 "rw_mbytes_per_sec": 0, 00:09:53.315 "r_mbytes_per_sec": 0, 00:09:53.315 "w_mbytes_per_sec": 0 00:09:53.315 }, 00:09:53.315 "claimed": true, 00:09:53.315 "claim_type": "exclusive_write", 00:09:53.315 "zoned": false, 00:09:53.315 "supported_io_types": { 00:09:53.315 "read": true, 00:09:53.315 "write": true, 00:09:53.315 "unmap": true, 00:09:53.315 "flush": true, 00:09:53.315 "reset": true, 00:09:53.315 "nvme_admin": false, 00:09:53.315 "nvme_io": false, 00:09:53.315 "nvme_io_md": false, 00:09:53.315 "write_zeroes": true, 00:09:53.315 "zcopy": true, 00:09:53.315 "get_zone_info": false, 00:09:53.315 "zone_management": false, 00:09:53.315 "zone_append": false, 00:09:53.315 "compare": false, 00:09:53.315 "compare_and_write": false, 00:09:53.315 "abort": true, 00:09:53.315 "seek_hole": false, 00:09:53.315 "seek_data": false, 00:09:53.315 "copy": true, 00:09:53.315 "nvme_iov_md": false 00:09:53.315 }, 00:09:53.315 "memory_domains": [ 00:09:53.315 { 00:09:53.315 "dma_device_id": "system", 00:09:53.315 "dma_device_type": 1 00:09:53.315 }, 00:09:53.315 { 00:09:53.315 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:53.315 "dma_device_type": 2 00:09:53.315 } 00:09:53.315 ], 00:09:53.315 "driver_specific": {} 00:09:53.315 } 00:09:53.315 ] 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.315 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.575 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:53.575 "name": "Existed_Raid", 00:09:53.575 "uuid": "6d349b67-4c01-45b0-8682-e884a89a3c55", 00:09:53.575 "strip_size_kb": 64, 00:09:53.575 "state": "online", 00:09:53.575 "raid_level": "concat", 00:09:53.575 "superblock": false, 00:09:53.575 "num_base_bdevs": 4, 00:09:53.575 "num_base_bdevs_discovered": 4, 00:09:53.575 "num_base_bdevs_operational": 4, 00:09:53.575 "base_bdevs_list": [ 00:09:53.575 { 00:09:53.575 "name": "BaseBdev1", 00:09:53.575 "uuid": "d020532a-baa2-4e5a-91cb-c82ac7c6b484", 00:09:53.575 "is_configured": true, 00:09:53.575 "data_offset": 0, 00:09:53.575 "data_size": 65536 00:09:53.575 }, 00:09:53.575 { 00:09:53.575 "name": "BaseBdev2", 00:09:53.575 "uuid": "734315f4-8aff-426c-b59d-65ef43126edb", 00:09:53.575 "is_configured": true, 00:09:53.575 "data_offset": 0, 00:09:53.575 "data_size": 65536 00:09:53.575 }, 00:09:53.575 { 00:09:53.575 "name": "BaseBdev3", 00:09:53.575 "uuid": "d6f56f82-d766-4364-ab4f-54f0326bcb42", 00:09:53.575 "is_configured": true, 00:09:53.575 "data_offset": 0, 00:09:53.575 "data_size": 65536 00:09:53.575 }, 00:09:53.575 { 00:09:53.575 "name": "BaseBdev4", 00:09:53.575 "uuid": "cd937bf8-0bb5-4626-bd77-415c929d75c9", 00:09:53.575 "is_configured": true, 00:09:53.575 "data_offset": 0, 00:09:53.575 "data_size": 65536 00:09:53.575 } 00:09:53.575 ] 00:09:53.575 }' 00:09:53.575 16:48:14 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:53.575 16:48:14 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.835 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:53.835 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:53.835 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:53.835 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:53.835 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.836 [2024-09-29 16:48:15.428120] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:53.836 "name": "Existed_Raid", 00:09:53.836 "aliases": [ 00:09:53.836 "6d349b67-4c01-45b0-8682-e884a89a3c55" 00:09:53.836 ], 00:09:53.836 "product_name": "Raid Volume", 00:09:53.836 "block_size": 512, 00:09:53.836 "num_blocks": 262144, 00:09:53.836 "uuid": "6d349b67-4c01-45b0-8682-e884a89a3c55", 00:09:53.836 "assigned_rate_limits": { 00:09:53.836 "rw_ios_per_sec": 0, 00:09:53.836 "rw_mbytes_per_sec": 0, 00:09:53.836 "r_mbytes_per_sec": 0, 00:09:53.836 "w_mbytes_per_sec": 0 00:09:53.836 }, 00:09:53.836 "claimed": false, 00:09:53.836 "zoned": false, 00:09:53.836 "supported_io_types": { 00:09:53.836 "read": true, 00:09:53.836 "write": true, 00:09:53.836 "unmap": true, 00:09:53.836 "flush": true, 00:09:53.836 "reset": true, 00:09:53.836 "nvme_admin": false, 00:09:53.836 "nvme_io": false, 00:09:53.836 "nvme_io_md": false, 00:09:53.836 "write_zeroes": true, 00:09:53.836 "zcopy": false, 00:09:53.836 "get_zone_info": false, 00:09:53.836 "zone_management": false, 00:09:53.836 "zone_append": false, 00:09:53.836 "compare": false, 00:09:53.836 "compare_and_write": false, 00:09:53.836 "abort": false, 00:09:53.836 "seek_hole": false, 00:09:53.836 "seek_data": false, 00:09:53.836 "copy": false, 00:09:53.836 "nvme_iov_md": false 00:09:53.836 }, 00:09:53.836 "memory_domains": [ 00:09:53.836 { 00:09:53.836 "dma_device_id": "system", 00:09:53.836 "dma_device_type": 1 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:53.836 "dma_device_type": 2 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "dma_device_id": "system", 00:09:53.836 "dma_device_type": 1 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:53.836 "dma_device_type": 2 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "dma_device_id": "system", 00:09:53.836 "dma_device_type": 1 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:53.836 "dma_device_type": 2 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "dma_device_id": "system", 00:09:53.836 "dma_device_type": 1 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:53.836 "dma_device_type": 2 00:09:53.836 } 00:09:53.836 ], 00:09:53.836 "driver_specific": { 00:09:53.836 "raid": { 00:09:53.836 "uuid": "6d349b67-4c01-45b0-8682-e884a89a3c55", 00:09:53.836 "strip_size_kb": 64, 00:09:53.836 "state": "online", 00:09:53.836 "raid_level": "concat", 00:09:53.836 "superblock": false, 00:09:53.836 "num_base_bdevs": 4, 00:09:53.836 "num_base_bdevs_discovered": 4, 00:09:53.836 "num_base_bdevs_operational": 4, 00:09:53.836 "base_bdevs_list": [ 00:09:53.836 { 00:09:53.836 "name": "BaseBdev1", 00:09:53.836 "uuid": "d020532a-baa2-4e5a-91cb-c82ac7c6b484", 00:09:53.836 "is_configured": true, 00:09:53.836 "data_offset": 0, 00:09:53.836 "data_size": 65536 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "name": "BaseBdev2", 00:09:53.836 "uuid": "734315f4-8aff-426c-b59d-65ef43126edb", 00:09:53.836 "is_configured": true, 00:09:53.836 "data_offset": 0, 00:09:53.836 "data_size": 65536 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "name": "BaseBdev3", 00:09:53.836 "uuid": "d6f56f82-d766-4364-ab4f-54f0326bcb42", 00:09:53.836 "is_configured": true, 00:09:53.836 "data_offset": 0, 00:09:53.836 "data_size": 65536 00:09:53.836 }, 00:09:53.836 { 00:09:53.836 "name": "BaseBdev4", 00:09:53.836 "uuid": "cd937bf8-0bb5-4626-bd77-415c929d75c9", 00:09:53.836 "is_configured": true, 00:09:53.836 "data_offset": 0, 00:09:53.836 "data_size": 65536 00:09:53.836 } 00:09:53.836 ] 00:09:53.836 } 00:09:53.836 } 00:09:53.836 }' 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:53.836 BaseBdev2 00:09:53.836 BaseBdev3 00:09:53.836 BaseBdev4' 00:09:53.836 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.097 [2024-09-29 16:48:15.723375] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:54.097 [2024-09-29 16:48:15.723444] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:54.097 [2024-09-29 16:48:15.723542] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.097 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.357 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.357 "name": "Existed_Raid", 00:09:54.357 "uuid": "6d349b67-4c01-45b0-8682-e884a89a3c55", 00:09:54.357 "strip_size_kb": 64, 00:09:54.357 "state": "offline", 00:09:54.357 "raid_level": "concat", 00:09:54.357 "superblock": false, 00:09:54.357 "num_base_bdevs": 4, 00:09:54.357 "num_base_bdevs_discovered": 3, 00:09:54.357 "num_base_bdevs_operational": 3, 00:09:54.357 "base_bdevs_list": [ 00:09:54.357 { 00:09:54.357 "name": null, 00:09:54.357 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:54.357 "is_configured": false, 00:09:54.357 "data_offset": 0, 00:09:54.357 "data_size": 65536 00:09:54.357 }, 00:09:54.357 { 00:09:54.357 "name": "BaseBdev2", 00:09:54.357 "uuid": "734315f4-8aff-426c-b59d-65ef43126edb", 00:09:54.357 "is_configured": true, 00:09:54.357 "data_offset": 0, 00:09:54.357 "data_size": 65536 00:09:54.357 }, 00:09:54.357 { 00:09:54.357 "name": "BaseBdev3", 00:09:54.357 "uuid": "d6f56f82-d766-4364-ab4f-54f0326bcb42", 00:09:54.357 "is_configured": true, 00:09:54.357 "data_offset": 0, 00:09:54.357 "data_size": 65536 00:09:54.357 }, 00:09:54.357 { 00:09:54.357 "name": "BaseBdev4", 00:09:54.357 "uuid": "cd937bf8-0bb5-4626-bd77-415c929d75c9", 00:09:54.357 "is_configured": true, 00:09:54.357 "data_offset": 0, 00:09:54.357 "data_size": 65536 00:09:54.357 } 00:09:54.357 ] 00:09:54.357 }' 00:09:54.357 16:48:15 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.357 16:48:15 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.618 [2024-09-29 16:48:16.166079] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.618 [2024-09-29 16:48:16.237112] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.618 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.879 [2024-09-29 16:48:16.308090] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:09:54.879 [2024-09-29 16:48:16.308189] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:54.879 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 BaseBdev2 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 [ 00:09:54.880 { 00:09:54.880 "name": "BaseBdev2", 00:09:54.880 "aliases": [ 00:09:54.880 "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc" 00:09:54.880 ], 00:09:54.880 "product_name": "Malloc disk", 00:09:54.880 "block_size": 512, 00:09:54.880 "num_blocks": 65536, 00:09:54.880 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:54.880 "assigned_rate_limits": { 00:09:54.880 "rw_ios_per_sec": 0, 00:09:54.880 "rw_mbytes_per_sec": 0, 00:09:54.880 "r_mbytes_per_sec": 0, 00:09:54.880 "w_mbytes_per_sec": 0 00:09:54.880 }, 00:09:54.880 "claimed": false, 00:09:54.880 "zoned": false, 00:09:54.880 "supported_io_types": { 00:09:54.880 "read": true, 00:09:54.880 "write": true, 00:09:54.880 "unmap": true, 00:09:54.880 "flush": true, 00:09:54.880 "reset": true, 00:09:54.880 "nvme_admin": false, 00:09:54.880 "nvme_io": false, 00:09:54.880 "nvme_io_md": false, 00:09:54.880 "write_zeroes": true, 00:09:54.880 "zcopy": true, 00:09:54.880 "get_zone_info": false, 00:09:54.880 "zone_management": false, 00:09:54.880 "zone_append": false, 00:09:54.880 "compare": false, 00:09:54.880 "compare_and_write": false, 00:09:54.880 "abort": true, 00:09:54.880 "seek_hole": false, 00:09:54.880 "seek_data": false, 00:09:54.880 "copy": true, 00:09:54.880 "nvme_iov_md": false 00:09:54.880 }, 00:09:54.880 "memory_domains": [ 00:09:54.880 { 00:09:54.880 "dma_device_id": "system", 00:09:54.880 "dma_device_type": 1 00:09:54.880 }, 00:09:54.880 { 00:09:54.880 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.880 "dma_device_type": 2 00:09:54.880 } 00:09:54.880 ], 00:09:54.880 "driver_specific": {} 00:09:54.880 } 00:09:54.880 ] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 BaseBdev3 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 [ 00:09:54.880 { 00:09:54.880 "name": "BaseBdev3", 00:09:54.880 "aliases": [ 00:09:54.880 "f8a49746-0cae-4576-b1cf-182f0737b616" 00:09:54.880 ], 00:09:54.880 "product_name": "Malloc disk", 00:09:54.880 "block_size": 512, 00:09:54.880 "num_blocks": 65536, 00:09:54.880 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:54.880 "assigned_rate_limits": { 00:09:54.880 "rw_ios_per_sec": 0, 00:09:54.880 "rw_mbytes_per_sec": 0, 00:09:54.880 "r_mbytes_per_sec": 0, 00:09:54.880 "w_mbytes_per_sec": 0 00:09:54.880 }, 00:09:54.880 "claimed": false, 00:09:54.880 "zoned": false, 00:09:54.880 "supported_io_types": { 00:09:54.880 "read": true, 00:09:54.880 "write": true, 00:09:54.880 "unmap": true, 00:09:54.880 "flush": true, 00:09:54.880 "reset": true, 00:09:54.880 "nvme_admin": false, 00:09:54.880 "nvme_io": false, 00:09:54.880 "nvme_io_md": false, 00:09:54.880 "write_zeroes": true, 00:09:54.880 "zcopy": true, 00:09:54.880 "get_zone_info": false, 00:09:54.880 "zone_management": false, 00:09:54.880 "zone_append": false, 00:09:54.880 "compare": false, 00:09:54.880 "compare_and_write": false, 00:09:54.880 "abort": true, 00:09:54.880 "seek_hole": false, 00:09:54.880 "seek_data": false, 00:09:54.880 "copy": true, 00:09:54.880 "nvme_iov_md": false 00:09:54.880 }, 00:09:54.880 "memory_domains": [ 00:09:54.880 { 00:09:54.880 "dma_device_id": "system", 00:09:54.880 "dma_device_type": 1 00:09:54.880 }, 00:09:54.880 { 00:09:54.880 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.880 "dma_device_type": 2 00:09:54.880 } 00:09:54.880 ], 00:09:54.880 "driver_specific": {} 00:09:54.880 } 00:09:54.880 ] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 BaseBdev4 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.880 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.880 [ 00:09:54.880 { 00:09:54.880 "name": "BaseBdev4", 00:09:54.880 "aliases": [ 00:09:54.880 "9bbe440a-f75b-4712-8e0b-f823ae8c0e12" 00:09:54.880 ], 00:09:54.880 "product_name": "Malloc disk", 00:09:54.880 "block_size": 512, 00:09:54.880 "num_blocks": 65536, 00:09:54.880 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:54.880 "assigned_rate_limits": { 00:09:54.880 "rw_ios_per_sec": 0, 00:09:54.880 "rw_mbytes_per_sec": 0, 00:09:54.881 "r_mbytes_per_sec": 0, 00:09:54.881 "w_mbytes_per_sec": 0 00:09:54.881 }, 00:09:54.881 "claimed": false, 00:09:54.881 "zoned": false, 00:09:54.881 "supported_io_types": { 00:09:54.881 "read": true, 00:09:54.881 "write": true, 00:09:54.881 "unmap": true, 00:09:54.881 "flush": true, 00:09:54.881 "reset": true, 00:09:54.881 "nvme_admin": false, 00:09:54.881 "nvme_io": false, 00:09:54.881 "nvme_io_md": false, 00:09:54.881 "write_zeroes": true, 00:09:54.881 "zcopy": true, 00:09:54.881 "get_zone_info": false, 00:09:54.881 "zone_management": false, 00:09:54.881 "zone_append": false, 00:09:54.881 "compare": false, 00:09:54.881 "compare_and_write": false, 00:09:54.881 "abort": true, 00:09:54.881 "seek_hole": false, 00:09:54.881 "seek_data": false, 00:09:54.881 "copy": true, 00:09:54.881 "nvme_iov_md": false 00:09:54.881 }, 00:09:54.881 "memory_domains": [ 00:09:54.881 { 00:09:54.881 "dma_device_id": "system", 00:09:54.881 "dma_device_type": 1 00:09:54.881 }, 00:09:54.881 { 00:09:54.881 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:54.881 "dma_device_type": 2 00:09:54.881 } 00:09:54.881 ], 00:09:54.881 "driver_specific": {} 00:09:54.881 } 00:09:54.881 ] 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.881 [2024-09-29 16:48:16.531337] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:54.881 [2024-09-29 16:48:16.531424] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:54.881 [2024-09-29 16:48:16.531465] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:54.881 [2024-09-29 16:48:16.533255] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:54.881 [2024-09-29 16:48:16.533351] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.881 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.141 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.141 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.141 "name": "Existed_Raid", 00:09:55.141 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.141 "strip_size_kb": 64, 00:09:55.141 "state": "configuring", 00:09:55.141 "raid_level": "concat", 00:09:55.141 "superblock": false, 00:09:55.141 "num_base_bdevs": 4, 00:09:55.141 "num_base_bdevs_discovered": 3, 00:09:55.141 "num_base_bdevs_operational": 4, 00:09:55.141 "base_bdevs_list": [ 00:09:55.141 { 00:09:55.141 "name": "BaseBdev1", 00:09:55.141 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.141 "is_configured": false, 00:09:55.141 "data_offset": 0, 00:09:55.141 "data_size": 0 00:09:55.141 }, 00:09:55.141 { 00:09:55.141 "name": "BaseBdev2", 00:09:55.141 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:55.141 "is_configured": true, 00:09:55.141 "data_offset": 0, 00:09:55.141 "data_size": 65536 00:09:55.141 }, 00:09:55.141 { 00:09:55.141 "name": "BaseBdev3", 00:09:55.141 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:55.141 "is_configured": true, 00:09:55.141 "data_offset": 0, 00:09:55.141 "data_size": 65536 00:09:55.141 }, 00:09:55.141 { 00:09:55.141 "name": "BaseBdev4", 00:09:55.141 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:55.141 "is_configured": true, 00:09:55.141 "data_offset": 0, 00:09:55.141 "data_size": 65536 00:09:55.141 } 00:09:55.141 ] 00:09:55.141 }' 00:09:55.141 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.141 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.402 [2024-09-29 16:48:16.954602] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.402 16:48:16 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.402 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.402 "name": "Existed_Raid", 00:09:55.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.402 "strip_size_kb": 64, 00:09:55.402 "state": "configuring", 00:09:55.402 "raid_level": "concat", 00:09:55.402 "superblock": false, 00:09:55.402 "num_base_bdevs": 4, 00:09:55.402 "num_base_bdevs_discovered": 2, 00:09:55.402 "num_base_bdevs_operational": 4, 00:09:55.402 "base_bdevs_list": [ 00:09:55.402 { 00:09:55.402 "name": "BaseBdev1", 00:09:55.402 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.402 "is_configured": false, 00:09:55.402 "data_offset": 0, 00:09:55.402 "data_size": 0 00:09:55.402 }, 00:09:55.402 { 00:09:55.402 "name": null, 00:09:55.402 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:55.402 "is_configured": false, 00:09:55.402 "data_offset": 0, 00:09:55.402 "data_size": 65536 00:09:55.402 }, 00:09:55.402 { 00:09:55.402 "name": "BaseBdev3", 00:09:55.402 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:55.402 "is_configured": true, 00:09:55.402 "data_offset": 0, 00:09:55.402 "data_size": 65536 00:09:55.402 }, 00:09:55.402 { 00:09:55.402 "name": "BaseBdev4", 00:09:55.402 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:55.402 "is_configured": true, 00:09:55.402 "data_offset": 0, 00:09:55.402 "data_size": 65536 00:09:55.402 } 00:09:55.402 ] 00:09:55.402 }' 00:09:55.402 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.402 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.972 [2024-09-29 16:48:17.424850] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:55.972 BaseBdev1 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.972 [ 00:09:55.972 { 00:09:55.972 "name": "BaseBdev1", 00:09:55.972 "aliases": [ 00:09:55.972 "62d5079b-a51b-4577-b422-4aa2008b7d3e" 00:09:55.972 ], 00:09:55.972 "product_name": "Malloc disk", 00:09:55.972 "block_size": 512, 00:09:55.972 "num_blocks": 65536, 00:09:55.972 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:55.972 "assigned_rate_limits": { 00:09:55.972 "rw_ios_per_sec": 0, 00:09:55.972 "rw_mbytes_per_sec": 0, 00:09:55.972 "r_mbytes_per_sec": 0, 00:09:55.972 "w_mbytes_per_sec": 0 00:09:55.972 }, 00:09:55.972 "claimed": true, 00:09:55.972 "claim_type": "exclusive_write", 00:09:55.972 "zoned": false, 00:09:55.972 "supported_io_types": { 00:09:55.972 "read": true, 00:09:55.972 "write": true, 00:09:55.972 "unmap": true, 00:09:55.972 "flush": true, 00:09:55.972 "reset": true, 00:09:55.972 "nvme_admin": false, 00:09:55.972 "nvme_io": false, 00:09:55.972 "nvme_io_md": false, 00:09:55.972 "write_zeroes": true, 00:09:55.972 "zcopy": true, 00:09:55.972 "get_zone_info": false, 00:09:55.972 "zone_management": false, 00:09:55.972 "zone_append": false, 00:09:55.972 "compare": false, 00:09:55.972 "compare_and_write": false, 00:09:55.972 "abort": true, 00:09:55.972 "seek_hole": false, 00:09:55.972 "seek_data": false, 00:09:55.972 "copy": true, 00:09:55.972 "nvme_iov_md": false 00:09:55.972 }, 00:09:55.972 "memory_domains": [ 00:09:55.972 { 00:09:55.972 "dma_device_id": "system", 00:09:55.972 "dma_device_type": 1 00:09:55.972 }, 00:09:55.972 { 00:09:55.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:55.972 "dma_device_type": 2 00:09:55.972 } 00:09:55.972 ], 00:09:55.972 "driver_specific": {} 00:09:55.972 } 00:09:55.972 ] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:55.972 "name": "Existed_Raid", 00:09:55.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:55.972 "strip_size_kb": 64, 00:09:55.972 "state": "configuring", 00:09:55.972 "raid_level": "concat", 00:09:55.972 "superblock": false, 00:09:55.972 "num_base_bdevs": 4, 00:09:55.972 "num_base_bdevs_discovered": 3, 00:09:55.972 "num_base_bdevs_operational": 4, 00:09:55.972 "base_bdevs_list": [ 00:09:55.972 { 00:09:55.972 "name": "BaseBdev1", 00:09:55.972 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:55.972 "is_configured": true, 00:09:55.972 "data_offset": 0, 00:09:55.972 "data_size": 65536 00:09:55.972 }, 00:09:55.972 { 00:09:55.972 "name": null, 00:09:55.972 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:55.972 "is_configured": false, 00:09:55.972 "data_offset": 0, 00:09:55.972 "data_size": 65536 00:09:55.972 }, 00:09:55.972 { 00:09:55.972 "name": "BaseBdev3", 00:09:55.972 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:55.972 "is_configured": true, 00:09:55.972 "data_offset": 0, 00:09:55.972 "data_size": 65536 00:09:55.972 }, 00:09:55.972 { 00:09:55.972 "name": "BaseBdev4", 00:09:55.972 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:55.972 "is_configured": true, 00:09:55.972 "data_offset": 0, 00:09:55.972 "data_size": 65536 00:09:55.972 } 00:09:55.972 ] 00:09:55.972 }' 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:55.972 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.232 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.232 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:56.232 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.232 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.491 [2024-09-29 16:48:17.940013] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.491 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.491 "name": "Existed_Raid", 00:09:56.491 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.491 "strip_size_kb": 64, 00:09:56.491 "state": "configuring", 00:09:56.491 "raid_level": "concat", 00:09:56.491 "superblock": false, 00:09:56.491 "num_base_bdevs": 4, 00:09:56.491 "num_base_bdevs_discovered": 2, 00:09:56.491 "num_base_bdevs_operational": 4, 00:09:56.491 "base_bdevs_list": [ 00:09:56.491 { 00:09:56.491 "name": "BaseBdev1", 00:09:56.491 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:56.491 "is_configured": true, 00:09:56.491 "data_offset": 0, 00:09:56.491 "data_size": 65536 00:09:56.491 }, 00:09:56.491 { 00:09:56.491 "name": null, 00:09:56.491 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:56.491 "is_configured": false, 00:09:56.491 "data_offset": 0, 00:09:56.491 "data_size": 65536 00:09:56.491 }, 00:09:56.491 { 00:09:56.491 "name": null, 00:09:56.491 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:56.491 "is_configured": false, 00:09:56.491 "data_offset": 0, 00:09:56.491 "data_size": 65536 00:09:56.491 }, 00:09:56.491 { 00:09:56.492 "name": "BaseBdev4", 00:09:56.492 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:56.492 "is_configured": true, 00:09:56.492 "data_offset": 0, 00:09:56.492 "data_size": 65536 00:09:56.492 } 00:09:56.492 ] 00:09:56.492 }' 00:09:56.492 16:48:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.492 16:48:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.751 [2024-09-29 16:48:18.415262] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.751 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.011 "name": "Existed_Raid", 00:09:57.011 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.011 "strip_size_kb": 64, 00:09:57.011 "state": "configuring", 00:09:57.011 "raid_level": "concat", 00:09:57.011 "superblock": false, 00:09:57.011 "num_base_bdevs": 4, 00:09:57.011 "num_base_bdevs_discovered": 3, 00:09:57.011 "num_base_bdevs_operational": 4, 00:09:57.011 "base_bdevs_list": [ 00:09:57.011 { 00:09:57.011 "name": "BaseBdev1", 00:09:57.011 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:57.011 "is_configured": true, 00:09:57.011 "data_offset": 0, 00:09:57.011 "data_size": 65536 00:09:57.011 }, 00:09:57.011 { 00:09:57.011 "name": null, 00:09:57.011 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:57.011 "is_configured": false, 00:09:57.011 "data_offset": 0, 00:09:57.011 "data_size": 65536 00:09:57.011 }, 00:09:57.011 { 00:09:57.011 "name": "BaseBdev3", 00:09:57.011 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:57.011 "is_configured": true, 00:09:57.011 "data_offset": 0, 00:09:57.011 "data_size": 65536 00:09:57.011 }, 00:09:57.011 { 00:09:57.011 "name": "BaseBdev4", 00:09:57.011 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:57.011 "is_configured": true, 00:09:57.011 "data_offset": 0, 00:09:57.011 "data_size": 65536 00:09:57.011 } 00:09:57.011 ] 00:09:57.011 }' 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.011 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.271 [2024-09-29 16:48:18.886607] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.271 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.530 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.530 "name": "Existed_Raid", 00:09:57.530 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.530 "strip_size_kb": 64, 00:09:57.530 "state": "configuring", 00:09:57.530 "raid_level": "concat", 00:09:57.530 "superblock": false, 00:09:57.530 "num_base_bdevs": 4, 00:09:57.530 "num_base_bdevs_discovered": 2, 00:09:57.530 "num_base_bdevs_operational": 4, 00:09:57.530 "base_bdevs_list": [ 00:09:57.530 { 00:09:57.530 "name": null, 00:09:57.530 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:57.530 "is_configured": false, 00:09:57.530 "data_offset": 0, 00:09:57.530 "data_size": 65536 00:09:57.530 }, 00:09:57.530 { 00:09:57.530 "name": null, 00:09:57.530 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:57.530 "is_configured": false, 00:09:57.530 "data_offset": 0, 00:09:57.530 "data_size": 65536 00:09:57.530 }, 00:09:57.530 { 00:09:57.530 "name": "BaseBdev3", 00:09:57.530 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:57.530 "is_configured": true, 00:09:57.530 "data_offset": 0, 00:09:57.530 "data_size": 65536 00:09:57.530 }, 00:09:57.530 { 00:09:57.530 "name": "BaseBdev4", 00:09:57.530 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:57.530 "is_configured": true, 00:09:57.530 "data_offset": 0, 00:09:57.530 "data_size": 65536 00:09:57.530 } 00:09:57.530 ] 00:09:57.530 }' 00:09:57.530 16:48:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.530 16:48:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.789 [2024-09-29 16:48:19.276481] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:57.789 "name": "Existed_Raid", 00:09:57.789 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:57.789 "strip_size_kb": 64, 00:09:57.789 "state": "configuring", 00:09:57.789 "raid_level": "concat", 00:09:57.789 "superblock": false, 00:09:57.789 "num_base_bdevs": 4, 00:09:57.789 "num_base_bdevs_discovered": 3, 00:09:57.789 "num_base_bdevs_operational": 4, 00:09:57.789 "base_bdevs_list": [ 00:09:57.789 { 00:09:57.789 "name": null, 00:09:57.789 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:57.789 "is_configured": false, 00:09:57.789 "data_offset": 0, 00:09:57.789 "data_size": 65536 00:09:57.789 }, 00:09:57.789 { 00:09:57.789 "name": "BaseBdev2", 00:09:57.789 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:57.789 "is_configured": true, 00:09:57.789 "data_offset": 0, 00:09:57.789 "data_size": 65536 00:09:57.789 }, 00:09:57.789 { 00:09:57.789 "name": "BaseBdev3", 00:09:57.789 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:57.789 "is_configured": true, 00:09:57.789 "data_offset": 0, 00:09:57.789 "data_size": 65536 00:09:57.789 }, 00:09:57.789 { 00:09:57.789 "name": "BaseBdev4", 00:09:57.789 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:57.789 "is_configured": true, 00:09:57.789 "data_offset": 0, 00:09:57.789 "data_size": 65536 00:09:57.789 } 00:09:57.789 ] 00:09:57.789 }' 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:57.789 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.049 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 62d5079b-a51b-4577-b422-4aa2008b7d3e 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.309 [2024-09-29 16:48:19.774597] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:58.309 [2024-09-29 16:48:19.774642] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:58.309 [2024-09-29 16:48:19.774649] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:09:58.309 [2024-09-29 16:48:19.774932] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:09:58.309 [2024-09-29 16:48:19.775078] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:58.309 [2024-09-29 16:48:19.775094] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:58.309 [2024-09-29 16:48:19.775292] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:58.309 NewBaseBdev 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:58.309 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.310 [ 00:09:58.310 { 00:09:58.310 "name": "NewBaseBdev", 00:09:58.310 "aliases": [ 00:09:58.310 "62d5079b-a51b-4577-b422-4aa2008b7d3e" 00:09:58.310 ], 00:09:58.310 "product_name": "Malloc disk", 00:09:58.310 "block_size": 512, 00:09:58.310 "num_blocks": 65536, 00:09:58.310 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:58.310 "assigned_rate_limits": { 00:09:58.310 "rw_ios_per_sec": 0, 00:09:58.310 "rw_mbytes_per_sec": 0, 00:09:58.310 "r_mbytes_per_sec": 0, 00:09:58.310 "w_mbytes_per_sec": 0 00:09:58.310 }, 00:09:58.310 "claimed": true, 00:09:58.310 "claim_type": "exclusive_write", 00:09:58.310 "zoned": false, 00:09:58.310 "supported_io_types": { 00:09:58.310 "read": true, 00:09:58.310 "write": true, 00:09:58.310 "unmap": true, 00:09:58.310 "flush": true, 00:09:58.310 "reset": true, 00:09:58.310 "nvme_admin": false, 00:09:58.310 "nvme_io": false, 00:09:58.310 "nvme_io_md": false, 00:09:58.310 "write_zeroes": true, 00:09:58.310 "zcopy": true, 00:09:58.310 "get_zone_info": false, 00:09:58.310 "zone_management": false, 00:09:58.310 "zone_append": false, 00:09:58.310 "compare": false, 00:09:58.310 "compare_and_write": false, 00:09:58.310 "abort": true, 00:09:58.310 "seek_hole": false, 00:09:58.310 "seek_data": false, 00:09:58.310 "copy": true, 00:09:58.310 "nvme_iov_md": false 00:09:58.310 }, 00:09:58.310 "memory_domains": [ 00:09:58.310 { 00:09:58.310 "dma_device_id": "system", 00:09:58.310 "dma_device_type": 1 00:09:58.310 }, 00:09:58.310 { 00:09:58.310 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.310 "dma_device_type": 2 00:09:58.310 } 00:09:58.310 ], 00:09:58.310 "driver_specific": {} 00:09:58.310 } 00:09:58.310 ] 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.310 "name": "Existed_Raid", 00:09:58.310 "uuid": "9a3217ab-6ccc-4781-b854-ed5031a6b93e", 00:09:58.310 "strip_size_kb": 64, 00:09:58.310 "state": "online", 00:09:58.310 "raid_level": "concat", 00:09:58.310 "superblock": false, 00:09:58.310 "num_base_bdevs": 4, 00:09:58.310 "num_base_bdevs_discovered": 4, 00:09:58.310 "num_base_bdevs_operational": 4, 00:09:58.310 "base_bdevs_list": [ 00:09:58.310 { 00:09:58.310 "name": "NewBaseBdev", 00:09:58.310 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:58.310 "is_configured": true, 00:09:58.310 "data_offset": 0, 00:09:58.310 "data_size": 65536 00:09:58.310 }, 00:09:58.310 { 00:09:58.310 "name": "BaseBdev2", 00:09:58.310 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:58.310 "is_configured": true, 00:09:58.310 "data_offset": 0, 00:09:58.310 "data_size": 65536 00:09:58.310 }, 00:09:58.310 { 00:09:58.310 "name": "BaseBdev3", 00:09:58.310 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:58.310 "is_configured": true, 00:09:58.310 "data_offset": 0, 00:09:58.310 "data_size": 65536 00:09:58.310 }, 00:09:58.310 { 00:09:58.310 "name": "BaseBdev4", 00:09:58.310 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:58.310 "is_configured": true, 00:09:58.310 "data_offset": 0, 00:09:58.310 "data_size": 65536 00:09:58.310 } 00:09:58.310 ] 00:09:58.310 }' 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.310 16:48:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:58.570 [2024-09-29 16:48:20.218168] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:58.570 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:58.831 "name": "Existed_Raid", 00:09:58.831 "aliases": [ 00:09:58.831 "9a3217ab-6ccc-4781-b854-ed5031a6b93e" 00:09:58.831 ], 00:09:58.831 "product_name": "Raid Volume", 00:09:58.831 "block_size": 512, 00:09:58.831 "num_blocks": 262144, 00:09:58.831 "uuid": "9a3217ab-6ccc-4781-b854-ed5031a6b93e", 00:09:58.831 "assigned_rate_limits": { 00:09:58.831 "rw_ios_per_sec": 0, 00:09:58.831 "rw_mbytes_per_sec": 0, 00:09:58.831 "r_mbytes_per_sec": 0, 00:09:58.831 "w_mbytes_per_sec": 0 00:09:58.831 }, 00:09:58.831 "claimed": false, 00:09:58.831 "zoned": false, 00:09:58.831 "supported_io_types": { 00:09:58.831 "read": true, 00:09:58.831 "write": true, 00:09:58.831 "unmap": true, 00:09:58.831 "flush": true, 00:09:58.831 "reset": true, 00:09:58.831 "nvme_admin": false, 00:09:58.831 "nvme_io": false, 00:09:58.831 "nvme_io_md": false, 00:09:58.831 "write_zeroes": true, 00:09:58.831 "zcopy": false, 00:09:58.831 "get_zone_info": false, 00:09:58.831 "zone_management": false, 00:09:58.831 "zone_append": false, 00:09:58.831 "compare": false, 00:09:58.831 "compare_and_write": false, 00:09:58.831 "abort": false, 00:09:58.831 "seek_hole": false, 00:09:58.831 "seek_data": false, 00:09:58.831 "copy": false, 00:09:58.831 "nvme_iov_md": false 00:09:58.831 }, 00:09:58.831 "memory_domains": [ 00:09:58.831 { 00:09:58.831 "dma_device_id": "system", 00:09:58.831 "dma_device_type": 1 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.831 "dma_device_type": 2 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "dma_device_id": "system", 00:09:58.831 "dma_device_type": 1 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.831 "dma_device_type": 2 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "dma_device_id": "system", 00:09:58.831 "dma_device_type": 1 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.831 "dma_device_type": 2 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "dma_device_id": "system", 00:09:58.831 "dma_device_type": 1 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.831 "dma_device_type": 2 00:09:58.831 } 00:09:58.831 ], 00:09:58.831 "driver_specific": { 00:09:58.831 "raid": { 00:09:58.831 "uuid": "9a3217ab-6ccc-4781-b854-ed5031a6b93e", 00:09:58.831 "strip_size_kb": 64, 00:09:58.831 "state": "online", 00:09:58.831 "raid_level": "concat", 00:09:58.831 "superblock": false, 00:09:58.831 "num_base_bdevs": 4, 00:09:58.831 "num_base_bdevs_discovered": 4, 00:09:58.831 "num_base_bdevs_operational": 4, 00:09:58.831 "base_bdevs_list": [ 00:09:58.831 { 00:09:58.831 "name": "NewBaseBdev", 00:09:58.831 "uuid": "62d5079b-a51b-4577-b422-4aa2008b7d3e", 00:09:58.831 "is_configured": true, 00:09:58.831 "data_offset": 0, 00:09:58.831 "data_size": 65536 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "name": "BaseBdev2", 00:09:58.831 "uuid": "6d8ab1cc-a320-4539-ba35-b17ae0bc1cbc", 00:09:58.831 "is_configured": true, 00:09:58.831 "data_offset": 0, 00:09:58.831 "data_size": 65536 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "name": "BaseBdev3", 00:09:58.831 "uuid": "f8a49746-0cae-4576-b1cf-182f0737b616", 00:09:58.831 "is_configured": true, 00:09:58.831 "data_offset": 0, 00:09:58.831 "data_size": 65536 00:09:58.831 }, 00:09:58.831 { 00:09:58.831 "name": "BaseBdev4", 00:09:58.831 "uuid": "9bbe440a-f75b-4712-8e0b-f823ae8c0e12", 00:09:58.831 "is_configured": true, 00:09:58.831 "data_offset": 0, 00:09:58.831 "data_size": 65536 00:09:58.831 } 00:09:58.831 ] 00:09:58.831 } 00:09:58.831 } 00:09:58.831 }' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:58.831 BaseBdev2 00:09:58.831 BaseBdev3 00:09:58.831 BaseBdev4' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.831 [2024-09-29 16:48:20.485398] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:58.831 [2024-09-29 16:48:20.485466] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:58.831 [2024-09-29 16:48:20.485560] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:58.831 [2024-09-29 16:48:20.485626] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:58.831 [2024-09-29 16:48:20.485636] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 81859 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 81859 ']' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 81859 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:58.831 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81859 00:09:59.091 killing process with pid 81859 00:09:59.091 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:59.091 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:59.091 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81859' 00:09:59.091 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 81859 00:09:59.091 [2024-09-29 16:48:20.517190] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:59.091 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 81859 00:09:59.091 [2024-09-29 16:48:20.556554] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:59.350 ************************************ 00:09:59.350 END TEST raid_state_function_test 00:09:59.350 ************************************ 00:09:59.350 00:09:59.350 real 0m9.167s 00:09:59.350 user 0m15.612s 00:09:59.350 sys 0m1.900s 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.350 16:48:20 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:09:59.350 16:48:20 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:59.350 16:48:20 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:59.350 16:48:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:59.350 ************************************ 00:09:59.350 START TEST raid_state_function_test_sb 00:09:59.350 ************************************ 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 true 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:59.350 Process raid pid: 82503 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82503 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82503' 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82503 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 82503 ']' 00:09:59.350 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:59.350 16:48:20 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:59.350 [2024-09-29 16:48:20.966259] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:09:59.350 [2024-09-29 16:48:20.966896] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:59.609 [2024-09-29 16:48:21.107298] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:59.609 [2024-09-29 16:48:21.154949] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:59.609 [2024-09-29 16:48:21.196902] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:59.609 [2024-09-29 16:48:21.197009] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:00.178 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:00.178 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.179 [2024-09-29 16:48:21.790119] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:00.179 [2024-09-29 16:48:21.790240] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:00.179 [2024-09-29 16:48:21.790272] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:00.179 [2024-09-29 16:48:21.790296] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:00.179 [2024-09-29 16:48:21.790313] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:00.179 [2024-09-29 16:48:21.790337] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:00.179 [2024-09-29 16:48:21.790354] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:00.179 [2024-09-29 16:48:21.790401] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.179 "name": "Existed_Raid", 00:10:00.179 "uuid": "a67bdd08-b328-4c00-a77d-c3cdb1effc12", 00:10:00.179 "strip_size_kb": 64, 00:10:00.179 "state": "configuring", 00:10:00.179 "raid_level": "concat", 00:10:00.179 "superblock": true, 00:10:00.179 "num_base_bdevs": 4, 00:10:00.179 "num_base_bdevs_discovered": 0, 00:10:00.179 "num_base_bdevs_operational": 4, 00:10:00.179 "base_bdevs_list": [ 00:10:00.179 { 00:10:00.179 "name": "BaseBdev1", 00:10:00.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.179 "is_configured": false, 00:10:00.179 "data_offset": 0, 00:10:00.179 "data_size": 0 00:10:00.179 }, 00:10:00.179 { 00:10:00.179 "name": "BaseBdev2", 00:10:00.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.179 "is_configured": false, 00:10:00.179 "data_offset": 0, 00:10:00.179 "data_size": 0 00:10:00.179 }, 00:10:00.179 { 00:10:00.179 "name": "BaseBdev3", 00:10:00.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.179 "is_configured": false, 00:10:00.179 "data_offset": 0, 00:10:00.179 "data_size": 0 00:10:00.179 }, 00:10:00.179 { 00:10:00.179 "name": "BaseBdev4", 00:10:00.179 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.179 "is_configured": false, 00:10:00.179 "data_offset": 0, 00:10:00.179 "data_size": 0 00:10:00.179 } 00:10:00.179 ] 00:10:00.179 }' 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.179 16:48:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.762 [2024-09-29 16:48:22.265222] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:00.762 [2024-09-29 16:48:22.265310] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.762 [2024-09-29 16:48:22.277220] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:00.762 [2024-09-29 16:48:22.277311] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:00.762 [2024-09-29 16:48:22.277338] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:00.762 [2024-09-29 16:48:22.277360] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:00.762 [2024-09-29 16:48:22.277377] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:00.762 [2024-09-29 16:48:22.277397] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:00.762 [2024-09-29 16:48:22.277415] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:00.762 [2024-09-29 16:48:22.277456] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.762 [2024-09-29 16:48:22.297964] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:00.762 BaseBdev1 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.762 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.762 [ 00:10:00.762 { 00:10:00.762 "name": "BaseBdev1", 00:10:00.762 "aliases": [ 00:10:00.762 "06ad3ac3-3929-475b-8e3d-1895983e8fce" 00:10:00.762 ], 00:10:00.762 "product_name": "Malloc disk", 00:10:00.762 "block_size": 512, 00:10:00.762 "num_blocks": 65536, 00:10:00.762 "uuid": "06ad3ac3-3929-475b-8e3d-1895983e8fce", 00:10:00.762 "assigned_rate_limits": { 00:10:00.762 "rw_ios_per_sec": 0, 00:10:00.762 "rw_mbytes_per_sec": 0, 00:10:00.762 "r_mbytes_per_sec": 0, 00:10:00.762 "w_mbytes_per_sec": 0 00:10:00.762 }, 00:10:00.762 "claimed": true, 00:10:00.762 "claim_type": "exclusive_write", 00:10:00.762 "zoned": false, 00:10:00.762 "supported_io_types": { 00:10:00.762 "read": true, 00:10:00.762 "write": true, 00:10:00.762 "unmap": true, 00:10:00.762 "flush": true, 00:10:00.762 "reset": true, 00:10:00.762 "nvme_admin": false, 00:10:00.762 "nvme_io": false, 00:10:00.762 "nvme_io_md": false, 00:10:00.763 "write_zeroes": true, 00:10:00.763 "zcopy": true, 00:10:00.763 "get_zone_info": false, 00:10:00.763 "zone_management": false, 00:10:00.763 "zone_append": false, 00:10:00.763 "compare": false, 00:10:00.763 "compare_and_write": false, 00:10:00.763 "abort": true, 00:10:00.763 "seek_hole": false, 00:10:00.763 "seek_data": false, 00:10:00.763 "copy": true, 00:10:00.763 "nvme_iov_md": false 00:10:00.763 }, 00:10:00.763 "memory_domains": [ 00:10:00.763 { 00:10:00.763 "dma_device_id": "system", 00:10:00.763 "dma_device_type": 1 00:10:00.763 }, 00:10:00.763 { 00:10:00.763 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.763 "dma_device_type": 2 00:10:00.763 } 00:10:00.763 ], 00:10:00.763 "driver_specific": {} 00:10:00.763 } 00:10:00.763 ] 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.763 "name": "Existed_Raid", 00:10:00.763 "uuid": "29b1a8f3-c6d7-4f9d-adef-0e407da8b171", 00:10:00.763 "strip_size_kb": 64, 00:10:00.763 "state": "configuring", 00:10:00.763 "raid_level": "concat", 00:10:00.763 "superblock": true, 00:10:00.763 "num_base_bdevs": 4, 00:10:00.763 "num_base_bdevs_discovered": 1, 00:10:00.763 "num_base_bdevs_operational": 4, 00:10:00.763 "base_bdevs_list": [ 00:10:00.763 { 00:10:00.763 "name": "BaseBdev1", 00:10:00.763 "uuid": "06ad3ac3-3929-475b-8e3d-1895983e8fce", 00:10:00.763 "is_configured": true, 00:10:00.763 "data_offset": 2048, 00:10:00.763 "data_size": 63488 00:10:00.763 }, 00:10:00.763 { 00:10:00.763 "name": "BaseBdev2", 00:10:00.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.763 "is_configured": false, 00:10:00.763 "data_offset": 0, 00:10:00.763 "data_size": 0 00:10:00.763 }, 00:10:00.763 { 00:10:00.763 "name": "BaseBdev3", 00:10:00.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.763 "is_configured": false, 00:10:00.763 "data_offset": 0, 00:10:00.763 "data_size": 0 00:10:00.763 }, 00:10:00.763 { 00:10:00.763 "name": "BaseBdev4", 00:10:00.763 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:00.763 "is_configured": false, 00:10:00.763 "data_offset": 0, 00:10:00.763 "data_size": 0 00:10:00.763 } 00:10:00.763 ] 00:10:00.763 }' 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.763 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.332 [2024-09-29 16:48:22.745246] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:01.332 [2024-09-29 16:48:22.745336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.332 [2024-09-29 16:48:22.757292] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:01.332 [2024-09-29 16:48:22.759196] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:01.332 [2024-09-29 16:48:22.759239] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:01.332 [2024-09-29 16:48:22.759249] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:01.332 [2024-09-29 16:48:22.759257] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:01.332 [2024-09-29 16:48:22.759264] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:01.332 [2024-09-29 16:48:22.759272] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.332 "name": "Existed_Raid", 00:10:01.332 "uuid": "853fb7c3-c441-43b5-a4be-f97ff7c1bb49", 00:10:01.332 "strip_size_kb": 64, 00:10:01.332 "state": "configuring", 00:10:01.332 "raid_level": "concat", 00:10:01.332 "superblock": true, 00:10:01.332 "num_base_bdevs": 4, 00:10:01.332 "num_base_bdevs_discovered": 1, 00:10:01.332 "num_base_bdevs_operational": 4, 00:10:01.332 "base_bdevs_list": [ 00:10:01.332 { 00:10:01.332 "name": "BaseBdev1", 00:10:01.332 "uuid": "06ad3ac3-3929-475b-8e3d-1895983e8fce", 00:10:01.332 "is_configured": true, 00:10:01.332 "data_offset": 2048, 00:10:01.332 "data_size": 63488 00:10:01.332 }, 00:10:01.332 { 00:10:01.332 "name": "BaseBdev2", 00:10:01.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.332 "is_configured": false, 00:10:01.332 "data_offset": 0, 00:10:01.332 "data_size": 0 00:10:01.332 }, 00:10:01.332 { 00:10:01.332 "name": "BaseBdev3", 00:10:01.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.332 "is_configured": false, 00:10:01.332 "data_offset": 0, 00:10:01.332 "data_size": 0 00:10:01.332 }, 00:10:01.332 { 00:10:01.332 "name": "BaseBdev4", 00:10:01.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.332 "is_configured": false, 00:10:01.332 "data_offset": 0, 00:10:01.332 "data_size": 0 00:10:01.332 } 00:10:01.332 ] 00:10:01.332 }' 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.332 16:48:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.592 [2024-09-29 16:48:23.166474] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:01.592 BaseBdev2 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.592 [ 00:10:01.592 { 00:10:01.592 "name": "BaseBdev2", 00:10:01.592 "aliases": [ 00:10:01.592 "cf642179-94ed-4842-ab3f-92f5edb180a7" 00:10:01.592 ], 00:10:01.592 "product_name": "Malloc disk", 00:10:01.592 "block_size": 512, 00:10:01.592 "num_blocks": 65536, 00:10:01.592 "uuid": "cf642179-94ed-4842-ab3f-92f5edb180a7", 00:10:01.592 "assigned_rate_limits": { 00:10:01.592 "rw_ios_per_sec": 0, 00:10:01.592 "rw_mbytes_per_sec": 0, 00:10:01.592 "r_mbytes_per_sec": 0, 00:10:01.592 "w_mbytes_per_sec": 0 00:10:01.592 }, 00:10:01.592 "claimed": true, 00:10:01.592 "claim_type": "exclusive_write", 00:10:01.592 "zoned": false, 00:10:01.592 "supported_io_types": { 00:10:01.592 "read": true, 00:10:01.592 "write": true, 00:10:01.592 "unmap": true, 00:10:01.592 "flush": true, 00:10:01.592 "reset": true, 00:10:01.592 "nvme_admin": false, 00:10:01.592 "nvme_io": false, 00:10:01.592 "nvme_io_md": false, 00:10:01.592 "write_zeroes": true, 00:10:01.592 "zcopy": true, 00:10:01.592 "get_zone_info": false, 00:10:01.592 "zone_management": false, 00:10:01.592 "zone_append": false, 00:10:01.592 "compare": false, 00:10:01.592 "compare_and_write": false, 00:10:01.592 "abort": true, 00:10:01.592 "seek_hole": false, 00:10:01.592 "seek_data": false, 00:10:01.592 "copy": true, 00:10:01.592 "nvme_iov_md": false 00:10:01.592 }, 00:10:01.592 "memory_domains": [ 00:10:01.592 { 00:10:01.592 "dma_device_id": "system", 00:10:01.592 "dma_device_type": 1 00:10:01.592 }, 00:10:01.592 { 00:10:01.592 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.592 "dma_device_type": 2 00:10:01.592 } 00:10:01.592 ], 00:10:01.592 "driver_specific": {} 00:10:01.592 } 00:10:01.592 ] 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.592 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.593 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.593 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.593 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.593 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:01.593 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.851 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.851 "name": "Existed_Raid", 00:10:01.851 "uuid": "853fb7c3-c441-43b5-a4be-f97ff7c1bb49", 00:10:01.851 "strip_size_kb": 64, 00:10:01.851 "state": "configuring", 00:10:01.851 "raid_level": "concat", 00:10:01.851 "superblock": true, 00:10:01.851 "num_base_bdevs": 4, 00:10:01.851 "num_base_bdevs_discovered": 2, 00:10:01.851 "num_base_bdevs_operational": 4, 00:10:01.851 "base_bdevs_list": [ 00:10:01.851 { 00:10:01.852 "name": "BaseBdev1", 00:10:01.852 "uuid": "06ad3ac3-3929-475b-8e3d-1895983e8fce", 00:10:01.852 "is_configured": true, 00:10:01.852 "data_offset": 2048, 00:10:01.852 "data_size": 63488 00:10:01.852 }, 00:10:01.852 { 00:10:01.852 "name": "BaseBdev2", 00:10:01.852 "uuid": "cf642179-94ed-4842-ab3f-92f5edb180a7", 00:10:01.852 "is_configured": true, 00:10:01.852 "data_offset": 2048, 00:10:01.852 "data_size": 63488 00:10:01.852 }, 00:10:01.852 { 00:10:01.852 "name": "BaseBdev3", 00:10:01.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.852 "is_configured": false, 00:10:01.852 "data_offset": 0, 00:10:01.852 "data_size": 0 00:10:01.852 }, 00:10:01.852 { 00:10:01.852 "name": "BaseBdev4", 00:10:01.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.852 "is_configured": false, 00:10:01.852 "data_offset": 0, 00:10:01.852 "data_size": 0 00:10:01.852 } 00:10:01.852 ] 00:10:01.852 }' 00:10:01.852 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.852 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.110 [2024-09-29 16:48:23.632587] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:02.110 BaseBdev3 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.110 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.110 [ 00:10:02.110 { 00:10:02.110 "name": "BaseBdev3", 00:10:02.110 "aliases": [ 00:10:02.111 "1bea2128-208c-4f20-aa72-973a8d50d2e4" 00:10:02.111 ], 00:10:02.111 "product_name": "Malloc disk", 00:10:02.111 "block_size": 512, 00:10:02.111 "num_blocks": 65536, 00:10:02.111 "uuid": "1bea2128-208c-4f20-aa72-973a8d50d2e4", 00:10:02.111 "assigned_rate_limits": { 00:10:02.111 "rw_ios_per_sec": 0, 00:10:02.111 "rw_mbytes_per_sec": 0, 00:10:02.111 "r_mbytes_per_sec": 0, 00:10:02.111 "w_mbytes_per_sec": 0 00:10:02.111 }, 00:10:02.111 "claimed": true, 00:10:02.111 "claim_type": "exclusive_write", 00:10:02.111 "zoned": false, 00:10:02.111 "supported_io_types": { 00:10:02.111 "read": true, 00:10:02.111 "write": true, 00:10:02.111 "unmap": true, 00:10:02.111 "flush": true, 00:10:02.111 "reset": true, 00:10:02.111 "nvme_admin": false, 00:10:02.111 "nvme_io": false, 00:10:02.111 "nvme_io_md": false, 00:10:02.111 "write_zeroes": true, 00:10:02.111 "zcopy": true, 00:10:02.111 "get_zone_info": false, 00:10:02.111 "zone_management": false, 00:10:02.111 "zone_append": false, 00:10:02.111 "compare": false, 00:10:02.111 "compare_and_write": false, 00:10:02.111 "abort": true, 00:10:02.111 "seek_hole": false, 00:10:02.111 "seek_data": false, 00:10:02.111 "copy": true, 00:10:02.111 "nvme_iov_md": false 00:10:02.111 }, 00:10:02.111 "memory_domains": [ 00:10:02.111 { 00:10:02.111 "dma_device_id": "system", 00:10:02.111 "dma_device_type": 1 00:10:02.111 }, 00:10:02.111 { 00:10:02.111 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.111 "dma_device_type": 2 00:10:02.111 } 00:10:02.111 ], 00:10:02.111 "driver_specific": {} 00:10:02.111 } 00:10:02.111 ] 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.111 "name": "Existed_Raid", 00:10:02.111 "uuid": "853fb7c3-c441-43b5-a4be-f97ff7c1bb49", 00:10:02.111 "strip_size_kb": 64, 00:10:02.111 "state": "configuring", 00:10:02.111 "raid_level": "concat", 00:10:02.111 "superblock": true, 00:10:02.111 "num_base_bdevs": 4, 00:10:02.111 "num_base_bdevs_discovered": 3, 00:10:02.111 "num_base_bdevs_operational": 4, 00:10:02.111 "base_bdevs_list": [ 00:10:02.111 { 00:10:02.111 "name": "BaseBdev1", 00:10:02.111 "uuid": "06ad3ac3-3929-475b-8e3d-1895983e8fce", 00:10:02.111 "is_configured": true, 00:10:02.111 "data_offset": 2048, 00:10:02.111 "data_size": 63488 00:10:02.111 }, 00:10:02.111 { 00:10:02.111 "name": "BaseBdev2", 00:10:02.111 "uuid": "cf642179-94ed-4842-ab3f-92f5edb180a7", 00:10:02.111 "is_configured": true, 00:10:02.111 "data_offset": 2048, 00:10:02.111 "data_size": 63488 00:10:02.111 }, 00:10:02.111 { 00:10:02.111 "name": "BaseBdev3", 00:10:02.111 "uuid": "1bea2128-208c-4f20-aa72-973a8d50d2e4", 00:10:02.111 "is_configured": true, 00:10:02.111 "data_offset": 2048, 00:10:02.111 "data_size": 63488 00:10:02.111 }, 00:10:02.111 { 00:10:02.111 "name": "BaseBdev4", 00:10:02.111 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.111 "is_configured": false, 00:10:02.111 "data_offset": 0, 00:10:02.111 "data_size": 0 00:10:02.111 } 00:10:02.111 ] 00:10:02.111 }' 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.111 16:48:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.679 BaseBdev4 00:10:02.679 [2024-09-29 16:48:24.146896] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:02.679 [2024-09-29 16:48:24.147134] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:02.679 [2024-09-29 16:48:24.147149] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:02.679 [2024-09-29 16:48:24.147426] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:02.679 [2024-09-29 16:48:24.147554] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:02.679 [2024-09-29 16:48:24.147579] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:02.679 [2024-09-29 16:48:24.147790] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.679 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.679 [ 00:10:02.679 { 00:10:02.679 "name": "BaseBdev4", 00:10:02.679 "aliases": [ 00:10:02.679 "cb325d7d-b9a7-4c94-8a20-b56b9d014656" 00:10:02.679 ], 00:10:02.679 "product_name": "Malloc disk", 00:10:02.679 "block_size": 512, 00:10:02.679 "num_blocks": 65536, 00:10:02.679 "uuid": "cb325d7d-b9a7-4c94-8a20-b56b9d014656", 00:10:02.679 "assigned_rate_limits": { 00:10:02.679 "rw_ios_per_sec": 0, 00:10:02.679 "rw_mbytes_per_sec": 0, 00:10:02.679 "r_mbytes_per_sec": 0, 00:10:02.679 "w_mbytes_per_sec": 0 00:10:02.679 }, 00:10:02.679 "claimed": true, 00:10:02.679 "claim_type": "exclusive_write", 00:10:02.679 "zoned": false, 00:10:02.679 "supported_io_types": { 00:10:02.679 "read": true, 00:10:02.679 "write": true, 00:10:02.679 "unmap": true, 00:10:02.679 "flush": true, 00:10:02.679 "reset": true, 00:10:02.679 "nvme_admin": false, 00:10:02.679 "nvme_io": false, 00:10:02.679 "nvme_io_md": false, 00:10:02.679 "write_zeroes": true, 00:10:02.679 "zcopy": true, 00:10:02.679 "get_zone_info": false, 00:10:02.679 "zone_management": false, 00:10:02.679 "zone_append": false, 00:10:02.679 "compare": false, 00:10:02.679 "compare_and_write": false, 00:10:02.679 "abort": true, 00:10:02.679 "seek_hole": false, 00:10:02.679 "seek_data": false, 00:10:02.679 "copy": true, 00:10:02.679 "nvme_iov_md": false 00:10:02.679 }, 00:10:02.680 "memory_domains": [ 00:10:02.680 { 00:10:02.680 "dma_device_id": "system", 00:10:02.680 "dma_device_type": 1 00:10:02.680 }, 00:10:02.680 { 00:10:02.680 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.680 "dma_device_type": 2 00:10:02.680 } 00:10:02.680 ], 00:10:02.680 "driver_specific": {} 00:10:02.680 } 00:10:02.680 ] 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.680 "name": "Existed_Raid", 00:10:02.680 "uuid": "853fb7c3-c441-43b5-a4be-f97ff7c1bb49", 00:10:02.680 "strip_size_kb": 64, 00:10:02.680 "state": "online", 00:10:02.680 "raid_level": "concat", 00:10:02.680 "superblock": true, 00:10:02.680 "num_base_bdevs": 4, 00:10:02.680 "num_base_bdevs_discovered": 4, 00:10:02.680 "num_base_bdevs_operational": 4, 00:10:02.680 "base_bdevs_list": [ 00:10:02.680 { 00:10:02.680 "name": "BaseBdev1", 00:10:02.680 "uuid": "06ad3ac3-3929-475b-8e3d-1895983e8fce", 00:10:02.680 "is_configured": true, 00:10:02.680 "data_offset": 2048, 00:10:02.680 "data_size": 63488 00:10:02.680 }, 00:10:02.680 { 00:10:02.680 "name": "BaseBdev2", 00:10:02.680 "uuid": "cf642179-94ed-4842-ab3f-92f5edb180a7", 00:10:02.680 "is_configured": true, 00:10:02.680 "data_offset": 2048, 00:10:02.680 "data_size": 63488 00:10:02.680 }, 00:10:02.680 { 00:10:02.680 "name": "BaseBdev3", 00:10:02.680 "uuid": "1bea2128-208c-4f20-aa72-973a8d50d2e4", 00:10:02.680 "is_configured": true, 00:10:02.680 "data_offset": 2048, 00:10:02.680 "data_size": 63488 00:10:02.680 }, 00:10:02.680 { 00:10:02.680 "name": "BaseBdev4", 00:10:02.680 "uuid": "cb325d7d-b9a7-4c94-8a20-b56b9d014656", 00:10:02.680 "is_configured": true, 00:10:02.680 "data_offset": 2048, 00:10:02.680 "data_size": 63488 00:10:02.680 } 00:10:02.680 ] 00:10:02.680 }' 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.680 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:02.939 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:03.198 [2024-09-29 16:48:24.614457] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:03.198 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.198 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:03.198 "name": "Existed_Raid", 00:10:03.198 "aliases": [ 00:10:03.198 "853fb7c3-c441-43b5-a4be-f97ff7c1bb49" 00:10:03.198 ], 00:10:03.198 "product_name": "Raid Volume", 00:10:03.198 "block_size": 512, 00:10:03.198 "num_blocks": 253952, 00:10:03.198 "uuid": "853fb7c3-c441-43b5-a4be-f97ff7c1bb49", 00:10:03.198 "assigned_rate_limits": { 00:10:03.198 "rw_ios_per_sec": 0, 00:10:03.198 "rw_mbytes_per_sec": 0, 00:10:03.198 "r_mbytes_per_sec": 0, 00:10:03.198 "w_mbytes_per_sec": 0 00:10:03.198 }, 00:10:03.198 "claimed": false, 00:10:03.198 "zoned": false, 00:10:03.198 "supported_io_types": { 00:10:03.198 "read": true, 00:10:03.198 "write": true, 00:10:03.198 "unmap": true, 00:10:03.198 "flush": true, 00:10:03.198 "reset": true, 00:10:03.198 "nvme_admin": false, 00:10:03.198 "nvme_io": false, 00:10:03.198 "nvme_io_md": false, 00:10:03.198 "write_zeroes": true, 00:10:03.198 "zcopy": false, 00:10:03.198 "get_zone_info": false, 00:10:03.198 "zone_management": false, 00:10:03.198 "zone_append": false, 00:10:03.198 "compare": false, 00:10:03.198 "compare_and_write": false, 00:10:03.198 "abort": false, 00:10:03.198 "seek_hole": false, 00:10:03.198 "seek_data": false, 00:10:03.198 "copy": false, 00:10:03.198 "nvme_iov_md": false 00:10:03.198 }, 00:10:03.198 "memory_domains": [ 00:10:03.198 { 00:10:03.198 "dma_device_id": "system", 00:10:03.198 "dma_device_type": 1 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.198 "dma_device_type": 2 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "dma_device_id": "system", 00:10:03.198 "dma_device_type": 1 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.198 "dma_device_type": 2 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "dma_device_id": "system", 00:10:03.198 "dma_device_type": 1 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.198 "dma_device_type": 2 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "dma_device_id": "system", 00:10:03.198 "dma_device_type": 1 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.198 "dma_device_type": 2 00:10:03.198 } 00:10:03.198 ], 00:10:03.198 "driver_specific": { 00:10:03.198 "raid": { 00:10:03.198 "uuid": "853fb7c3-c441-43b5-a4be-f97ff7c1bb49", 00:10:03.198 "strip_size_kb": 64, 00:10:03.198 "state": "online", 00:10:03.198 "raid_level": "concat", 00:10:03.198 "superblock": true, 00:10:03.198 "num_base_bdevs": 4, 00:10:03.198 "num_base_bdevs_discovered": 4, 00:10:03.198 "num_base_bdevs_operational": 4, 00:10:03.198 "base_bdevs_list": [ 00:10:03.198 { 00:10:03.198 "name": "BaseBdev1", 00:10:03.198 "uuid": "06ad3ac3-3929-475b-8e3d-1895983e8fce", 00:10:03.198 "is_configured": true, 00:10:03.198 "data_offset": 2048, 00:10:03.198 "data_size": 63488 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "name": "BaseBdev2", 00:10:03.198 "uuid": "cf642179-94ed-4842-ab3f-92f5edb180a7", 00:10:03.198 "is_configured": true, 00:10:03.198 "data_offset": 2048, 00:10:03.198 "data_size": 63488 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "name": "BaseBdev3", 00:10:03.198 "uuid": "1bea2128-208c-4f20-aa72-973a8d50d2e4", 00:10:03.198 "is_configured": true, 00:10:03.198 "data_offset": 2048, 00:10:03.198 "data_size": 63488 00:10:03.198 }, 00:10:03.198 { 00:10:03.198 "name": "BaseBdev4", 00:10:03.199 "uuid": "cb325d7d-b9a7-4c94-8a20-b56b9d014656", 00:10:03.199 "is_configured": true, 00:10:03.199 "data_offset": 2048, 00:10:03.199 "data_size": 63488 00:10:03.199 } 00:10:03.199 ] 00:10:03.199 } 00:10:03.199 } 00:10:03.199 }' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:03.199 BaseBdev2 00:10:03.199 BaseBdev3 00:10:03.199 BaseBdev4' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.199 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:03.458 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.458 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:03.458 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:03.458 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:03.458 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.458 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.458 [2024-09-29 16:48:24.917628] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:03.458 [2024-09-29 16:48:24.917657] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:03.458 [2024-09-29 16:48:24.917701] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:03.458 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.459 "name": "Existed_Raid", 00:10:03.459 "uuid": "853fb7c3-c441-43b5-a4be-f97ff7c1bb49", 00:10:03.459 "strip_size_kb": 64, 00:10:03.459 "state": "offline", 00:10:03.459 "raid_level": "concat", 00:10:03.459 "superblock": true, 00:10:03.459 "num_base_bdevs": 4, 00:10:03.459 "num_base_bdevs_discovered": 3, 00:10:03.459 "num_base_bdevs_operational": 3, 00:10:03.459 "base_bdevs_list": [ 00:10:03.459 { 00:10:03.459 "name": null, 00:10:03.459 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.459 "is_configured": false, 00:10:03.459 "data_offset": 0, 00:10:03.459 "data_size": 63488 00:10:03.459 }, 00:10:03.459 { 00:10:03.459 "name": "BaseBdev2", 00:10:03.459 "uuid": "cf642179-94ed-4842-ab3f-92f5edb180a7", 00:10:03.459 "is_configured": true, 00:10:03.459 "data_offset": 2048, 00:10:03.459 "data_size": 63488 00:10:03.459 }, 00:10:03.459 { 00:10:03.459 "name": "BaseBdev3", 00:10:03.459 "uuid": "1bea2128-208c-4f20-aa72-973a8d50d2e4", 00:10:03.459 "is_configured": true, 00:10:03.459 "data_offset": 2048, 00:10:03.459 "data_size": 63488 00:10:03.459 }, 00:10:03.459 { 00:10:03.459 "name": "BaseBdev4", 00:10:03.459 "uuid": "cb325d7d-b9a7-4c94-8a20-b56b9d014656", 00:10:03.459 "is_configured": true, 00:10:03.459 "data_offset": 2048, 00:10:03.459 "data_size": 63488 00:10:03.459 } 00:10:03.459 ] 00:10:03.459 }' 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.459 16:48:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.718 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:03.718 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:03.718 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.718 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:03.719 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.719 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 [2024-09-29 16:48:25.432219] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 [2024-09-29 16:48:25.487390] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 [2024-09-29 16:48:25.554525] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:03.979 [2024-09-29 16:48:25.554609] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 BaseBdev2 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.979 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.239 [ 00:10:04.239 { 00:10:04.239 "name": "BaseBdev2", 00:10:04.239 "aliases": [ 00:10:04.239 "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0" 00:10:04.239 ], 00:10:04.239 "product_name": "Malloc disk", 00:10:04.239 "block_size": 512, 00:10:04.239 "num_blocks": 65536, 00:10:04.239 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:04.239 "assigned_rate_limits": { 00:10:04.239 "rw_ios_per_sec": 0, 00:10:04.239 "rw_mbytes_per_sec": 0, 00:10:04.239 "r_mbytes_per_sec": 0, 00:10:04.239 "w_mbytes_per_sec": 0 00:10:04.239 }, 00:10:04.239 "claimed": false, 00:10:04.239 "zoned": false, 00:10:04.239 "supported_io_types": { 00:10:04.239 "read": true, 00:10:04.239 "write": true, 00:10:04.239 "unmap": true, 00:10:04.239 "flush": true, 00:10:04.239 "reset": true, 00:10:04.239 "nvme_admin": false, 00:10:04.239 "nvme_io": false, 00:10:04.239 "nvme_io_md": false, 00:10:04.239 "write_zeroes": true, 00:10:04.239 "zcopy": true, 00:10:04.239 "get_zone_info": false, 00:10:04.239 "zone_management": false, 00:10:04.239 "zone_append": false, 00:10:04.239 "compare": false, 00:10:04.239 "compare_and_write": false, 00:10:04.239 "abort": true, 00:10:04.239 "seek_hole": false, 00:10:04.239 "seek_data": false, 00:10:04.239 "copy": true, 00:10:04.239 "nvme_iov_md": false 00:10:04.239 }, 00:10:04.239 "memory_domains": [ 00:10:04.239 { 00:10:04.239 "dma_device_id": "system", 00:10:04.239 "dma_device_type": 1 00:10:04.239 }, 00:10:04.239 { 00:10:04.239 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.239 "dma_device_type": 2 00:10:04.239 } 00:10:04.239 ], 00:10:04.239 "driver_specific": {} 00:10:04.239 } 00:10:04.239 ] 00:10:04.239 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.239 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:04.239 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 BaseBdev3 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 [ 00:10:04.240 { 00:10:04.240 "name": "BaseBdev3", 00:10:04.240 "aliases": [ 00:10:04.240 "900d49d9-ac67-498d-b70d-347289c28a9c" 00:10:04.240 ], 00:10:04.240 "product_name": "Malloc disk", 00:10:04.240 "block_size": 512, 00:10:04.240 "num_blocks": 65536, 00:10:04.240 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:04.240 "assigned_rate_limits": { 00:10:04.240 "rw_ios_per_sec": 0, 00:10:04.240 "rw_mbytes_per_sec": 0, 00:10:04.240 "r_mbytes_per_sec": 0, 00:10:04.240 "w_mbytes_per_sec": 0 00:10:04.240 }, 00:10:04.240 "claimed": false, 00:10:04.240 "zoned": false, 00:10:04.240 "supported_io_types": { 00:10:04.240 "read": true, 00:10:04.240 "write": true, 00:10:04.240 "unmap": true, 00:10:04.240 "flush": true, 00:10:04.240 "reset": true, 00:10:04.240 "nvme_admin": false, 00:10:04.240 "nvme_io": false, 00:10:04.240 "nvme_io_md": false, 00:10:04.240 "write_zeroes": true, 00:10:04.240 "zcopy": true, 00:10:04.240 "get_zone_info": false, 00:10:04.240 "zone_management": false, 00:10:04.240 "zone_append": false, 00:10:04.240 "compare": false, 00:10:04.240 "compare_and_write": false, 00:10:04.240 "abort": true, 00:10:04.240 "seek_hole": false, 00:10:04.240 "seek_data": false, 00:10:04.240 "copy": true, 00:10:04.240 "nvme_iov_md": false 00:10:04.240 }, 00:10:04.240 "memory_domains": [ 00:10:04.240 { 00:10:04.240 "dma_device_id": "system", 00:10:04.240 "dma_device_type": 1 00:10:04.240 }, 00:10:04.240 { 00:10:04.240 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.240 "dma_device_type": 2 00:10:04.240 } 00:10:04.240 ], 00:10:04.240 "driver_specific": {} 00:10:04.240 } 00:10:04.240 ] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 BaseBdev4 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 [ 00:10:04.240 { 00:10:04.240 "name": "BaseBdev4", 00:10:04.240 "aliases": [ 00:10:04.240 "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323" 00:10:04.240 ], 00:10:04.240 "product_name": "Malloc disk", 00:10:04.240 "block_size": 512, 00:10:04.240 "num_blocks": 65536, 00:10:04.240 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:04.240 "assigned_rate_limits": { 00:10:04.240 "rw_ios_per_sec": 0, 00:10:04.240 "rw_mbytes_per_sec": 0, 00:10:04.240 "r_mbytes_per_sec": 0, 00:10:04.240 "w_mbytes_per_sec": 0 00:10:04.240 }, 00:10:04.240 "claimed": false, 00:10:04.240 "zoned": false, 00:10:04.240 "supported_io_types": { 00:10:04.240 "read": true, 00:10:04.240 "write": true, 00:10:04.240 "unmap": true, 00:10:04.240 "flush": true, 00:10:04.240 "reset": true, 00:10:04.240 "nvme_admin": false, 00:10:04.240 "nvme_io": false, 00:10:04.240 "nvme_io_md": false, 00:10:04.240 "write_zeroes": true, 00:10:04.240 "zcopy": true, 00:10:04.240 "get_zone_info": false, 00:10:04.240 "zone_management": false, 00:10:04.240 "zone_append": false, 00:10:04.240 "compare": false, 00:10:04.240 "compare_and_write": false, 00:10:04.240 "abort": true, 00:10:04.240 "seek_hole": false, 00:10:04.240 "seek_data": false, 00:10:04.240 "copy": true, 00:10:04.240 "nvme_iov_md": false 00:10:04.240 }, 00:10:04.240 "memory_domains": [ 00:10:04.240 { 00:10:04.240 "dma_device_id": "system", 00:10:04.240 "dma_device_type": 1 00:10:04.240 }, 00:10:04.240 { 00:10:04.240 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:04.240 "dma_device_type": 2 00:10:04.240 } 00:10:04.240 ], 00:10:04.240 "driver_specific": {} 00:10:04.240 } 00:10:04.240 ] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 [2024-09-29 16:48:25.785816] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:04.240 [2024-09-29 16:48:25.785909] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:04.240 [2024-09-29 16:48:25.785934] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:04.240 [2024-09-29 16:48:25.787773] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:04.240 [2024-09-29 16:48:25.787821] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.240 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.241 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.241 "name": "Existed_Raid", 00:10:04.241 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:04.241 "strip_size_kb": 64, 00:10:04.241 "state": "configuring", 00:10:04.241 "raid_level": "concat", 00:10:04.241 "superblock": true, 00:10:04.241 "num_base_bdevs": 4, 00:10:04.241 "num_base_bdevs_discovered": 3, 00:10:04.241 "num_base_bdevs_operational": 4, 00:10:04.241 "base_bdevs_list": [ 00:10:04.241 { 00:10:04.241 "name": "BaseBdev1", 00:10:04.241 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.241 "is_configured": false, 00:10:04.241 "data_offset": 0, 00:10:04.241 "data_size": 0 00:10:04.241 }, 00:10:04.241 { 00:10:04.241 "name": "BaseBdev2", 00:10:04.241 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:04.241 "is_configured": true, 00:10:04.241 "data_offset": 2048, 00:10:04.241 "data_size": 63488 00:10:04.241 }, 00:10:04.241 { 00:10:04.241 "name": "BaseBdev3", 00:10:04.241 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:04.241 "is_configured": true, 00:10:04.241 "data_offset": 2048, 00:10:04.241 "data_size": 63488 00:10:04.241 }, 00:10:04.241 { 00:10:04.241 "name": "BaseBdev4", 00:10:04.241 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:04.241 "is_configured": true, 00:10:04.241 "data_offset": 2048, 00:10:04.241 "data_size": 63488 00:10:04.241 } 00:10:04.241 ] 00:10:04.241 }' 00:10:04.241 16:48:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.241 16:48:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.810 [2024-09-29 16:48:26.197068] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.810 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.810 "name": "Existed_Raid", 00:10:04.810 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:04.810 "strip_size_kb": 64, 00:10:04.810 "state": "configuring", 00:10:04.810 "raid_level": "concat", 00:10:04.810 "superblock": true, 00:10:04.810 "num_base_bdevs": 4, 00:10:04.810 "num_base_bdevs_discovered": 2, 00:10:04.810 "num_base_bdevs_operational": 4, 00:10:04.810 "base_bdevs_list": [ 00:10:04.810 { 00:10:04.810 "name": "BaseBdev1", 00:10:04.810 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.810 "is_configured": false, 00:10:04.810 "data_offset": 0, 00:10:04.810 "data_size": 0 00:10:04.810 }, 00:10:04.810 { 00:10:04.810 "name": null, 00:10:04.810 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:04.810 "is_configured": false, 00:10:04.810 "data_offset": 0, 00:10:04.810 "data_size": 63488 00:10:04.810 }, 00:10:04.810 { 00:10:04.810 "name": "BaseBdev3", 00:10:04.810 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:04.810 "is_configured": true, 00:10:04.810 "data_offset": 2048, 00:10:04.810 "data_size": 63488 00:10:04.810 }, 00:10:04.810 { 00:10:04.810 "name": "BaseBdev4", 00:10:04.810 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:04.810 "is_configured": true, 00:10:04.811 "data_offset": 2048, 00:10:04.811 "data_size": 63488 00:10:04.811 } 00:10:04.811 ] 00:10:04.811 }' 00:10:04.811 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.811 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.070 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.070 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.071 [2024-09-29 16:48:26.663149] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:05.071 BaseBdev1 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.071 [ 00:10:05.071 { 00:10:05.071 "name": "BaseBdev1", 00:10:05.071 "aliases": [ 00:10:05.071 "28b77698-015a-4499-aa2e-1e0dd87baa52" 00:10:05.071 ], 00:10:05.071 "product_name": "Malloc disk", 00:10:05.071 "block_size": 512, 00:10:05.071 "num_blocks": 65536, 00:10:05.071 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:05.071 "assigned_rate_limits": { 00:10:05.071 "rw_ios_per_sec": 0, 00:10:05.071 "rw_mbytes_per_sec": 0, 00:10:05.071 "r_mbytes_per_sec": 0, 00:10:05.071 "w_mbytes_per_sec": 0 00:10:05.071 }, 00:10:05.071 "claimed": true, 00:10:05.071 "claim_type": "exclusive_write", 00:10:05.071 "zoned": false, 00:10:05.071 "supported_io_types": { 00:10:05.071 "read": true, 00:10:05.071 "write": true, 00:10:05.071 "unmap": true, 00:10:05.071 "flush": true, 00:10:05.071 "reset": true, 00:10:05.071 "nvme_admin": false, 00:10:05.071 "nvme_io": false, 00:10:05.071 "nvme_io_md": false, 00:10:05.071 "write_zeroes": true, 00:10:05.071 "zcopy": true, 00:10:05.071 "get_zone_info": false, 00:10:05.071 "zone_management": false, 00:10:05.071 "zone_append": false, 00:10:05.071 "compare": false, 00:10:05.071 "compare_and_write": false, 00:10:05.071 "abort": true, 00:10:05.071 "seek_hole": false, 00:10:05.071 "seek_data": false, 00:10:05.071 "copy": true, 00:10:05.071 "nvme_iov_md": false 00:10:05.071 }, 00:10:05.071 "memory_domains": [ 00:10:05.071 { 00:10:05.071 "dma_device_id": "system", 00:10:05.071 "dma_device_type": 1 00:10:05.071 }, 00:10:05.071 { 00:10:05.071 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.071 "dma_device_type": 2 00:10:05.071 } 00:10:05.071 ], 00:10:05.071 "driver_specific": {} 00:10:05.071 } 00:10:05.071 ] 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.071 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.330 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.330 "name": "Existed_Raid", 00:10:05.330 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:05.330 "strip_size_kb": 64, 00:10:05.330 "state": "configuring", 00:10:05.330 "raid_level": "concat", 00:10:05.330 "superblock": true, 00:10:05.330 "num_base_bdevs": 4, 00:10:05.330 "num_base_bdevs_discovered": 3, 00:10:05.330 "num_base_bdevs_operational": 4, 00:10:05.330 "base_bdevs_list": [ 00:10:05.330 { 00:10:05.330 "name": "BaseBdev1", 00:10:05.330 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:05.330 "is_configured": true, 00:10:05.330 "data_offset": 2048, 00:10:05.330 "data_size": 63488 00:10:05.330 }, 00:10:05.330 { 00:10:05.330 "name": null, 00:10:05.330 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:05.330 "is_configured": false, 00:10:05.330 "data_offset": 0, 00:10:05.330 "data_size": 63488 00:10:05.330 }, 00:10:05.330 { 00:10:05.330 "name": "BaseBdev3", 00:10:05.330 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:05.330 "is_configured": true, 00:10:05.330 "data_offset": 2048, 00:10:05.330 "data_size": 63488 00:10:05.330 }, 00:10:05.330 { 00:10:05.330 "name": "BaseBdev4", 00:10:05.330 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:05.330 "is_configured": true, 00:10:05.330 "data_offset": 2048, 00:10:05.330 "data_size": 63488 00:10:05.330 } 00:10:05.330 ] 00:10:05.330 }' 00:10:05.330 16:48:26 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.330 16:48:26 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.590 [2024-09-29 16:48:27.118405] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.590 "name": "Existed_Raid", 00:10:05.590 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:05.590 "strip_size_kb": 64, 00:10:05.590 "state": "configuring", 00:10:05.590 "raid_level": "concat", 00:10:05.590 "superblock": true, 00:10:05.590 "num_base_bdevs": 4, 00:10:05.590 "num_base_bdevs_discovered": 2, 00:10:05.590 "num_base_bdevs_operational": 4, 00:10:05.590 "base_bdevs_list": [ 00:10:05.590 { 00:10:05.590 "name": "BaseBdev1", 00:10:05.590 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:05.590 "is_configured": true, 00:10:05.590 "data_offset": 2048, 00:10:05.590 "data_size": 63488 00:10:05.590 }, 00:10:05.590 { 00:10:05.590 "name": null, 00:10:05.590 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:05.590 "is_configured": false, 00:10:05.590 "data_offset": 0, 00:10:05.590 "data_size": 63488 00:10:05.590 }, 00:10:05.590 { 00:10:05.590 "name": null, 00:10:05.590 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:05.590 "is_configured": false, 00:10:05.590 "data_offset": 0, 00:10:05.590 "data_size": 63488 00:10:05.590 }, 00:10:05.590 { 00:10:05.590 "name": "BaseBdev4", 00:10:05.590 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:05.590 "is_configured": true, 00:10:05.590 "data_offset": 2048, 00:10:05.590 "data_size": 63488 00:10:05.590 } 00:10:05.590 ] 00:10:05.590 }' 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.590 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.159 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.160 [2024-09-29 16:48:27.613621] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.160 "name": "Existed_Raid", 00:10:06.160 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:06.160 "strip_size_kb": 64, 00:10:06.160 "state": "configuring", 00:10:06.160 "raid_level": "concat", 00:10:06.160 "superblock": true, 00:10:06.160 "num_base_bdevs": 4, 00:10:06.160 "num_base_bdevs_discovered": 3, 00:10:06.160 "num_base_bdevs_operational": 4, 00:10:06.160 "base_bdevs_list": [ 00:10:06.160 { 00:10:06.160 "name": "BaseBdev1", 00:10:06.160 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:06.160 "is_configured": true, 00:10:06.160 "data_offset": 2048, 00:10:06.160 "data_size": 63488 00:10:06.160 }, 00:10:06.160 { 00:10:06.160 "name": null, 00:10:06.160 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:06.160 "is_configured": false, 00:10:06.160 "data_offset": 0, 00:10:06.160 "data_size": 63488 00:10:06.160 }, 00:10:06.160 { 00:10:06.160 "name": "BaseBdev3", 00:10:06.160 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:06.160 "is_configured": true, 00:10:06.160 "data_offset": 2048, 00:10:06.160 "data_size": 63488 00:10:06.160 }, 00:10:06.160 { 00:10:06.160 "name": "BaseBdev4", 00:10:06.160 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:06.160 "is_configured": true, 00:10:06.160 "data_offset": 2048, 00:10:06.160 "data_size": 63488 00:10:06.160 } 00:10:06.160 ] 00:10:06.160 }' 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.160 16:48:27 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.420 [2024-09-29 16:48:28.052885] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.420 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.680 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.680 "name": "Existed_Raid", 00:10:06.680 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:06.680 "strip_size_kb": 64, 00:10:06.680 "state": "configuring", 00:10:06.680 "raid_level": "concat", 00:10:06.680 "superblock": true, 00:10:06.680 "num_base_bdevs": 4, 00:10:06.680 "num_base_bdevs_discovered": 2, 00:10:06.680 "num_base_bdevs_operational": 4, 00:10:06.680 "base_bdevs_list": [ 00:10:06.680 { 00:10:06.680 "name": null, 00:10:06.680 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:06.680 "is_configured": false, 00:10:06.680 "data_offset": 0, 00:10:06.680 "data_size": 63488 00:10:06.680 }, 00:10:06.680 { 00:10:06.680 "name": null, 00:10:06.680 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:06.680 "is_configured": false, 00:10:06.680 "data_offset": 0, 00:10:06.680 "data_size": 63488 00:10:06.680 }, 00:10:06.680 { 00:10:06.680 "name": "BaseBdev3", 00:10:06.680 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:06.680 "is_configured": true, 00:10:06.680 "data_offset": 2048, 00:10:06.680 "data_size": 63488 00:10:06.680 }, 00:10:06.680 { 00:10:06.680 "name": "BaseBdev4", 00:10:06.680 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:06.680 "is_configured": true, 00:10:06.680 "data_offset": 2048, 00:10:06.680 "data_size": 63488 00:10:06.680 } 00:10:06.680 ] 00:10:06.680 }' 00:10:06.680 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.680 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.940 [2024-09-29 16:48:28.538625] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.940 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:06.940 "name": "Existed_Raid", 00:10:06.940 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:06.940 "strip_size_kb": 64, 00:10:06.940 "state": "configuring", 00:10:06.940 "raid_level": "concat", 00:10:06.940 "superblock": true, 00:10:06.940 "num_base_bdevs": 4, 00:10:06.940 "num_base_bdevs_discovered": 3, 00:10:06.940 "num_base_bdevs_operational": 4, 00:10:06.940 "base_bdevs_list": [ 00:10:06.940 { 00:10:06.940 "name": null, 00:10:06.940 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:06.940 "is_configured": false, 00:10:06.940 "data_offset": 0, 00:10:06.940 "data_size": 63488 00:10:06.940 }, 00:10:06.940 { 00:10:06.940 "name": "BaseBdev2", 00:10:06.940 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:06.940 "is_configured": true, 00:10:06.940 "data_offset": 2048, 00:10:06.940 "data_size": 63488 00:10:06.940 }, 00:10:06.940 { 00:10:06.940 "name": "BaseBdev3", 00:10:06.940 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:06.940 "is_configured": true, 00:10:06.941 "data_offset": 2048, 00:10:06.941 "data_size": 63488 00:10:06.941 }, 00:10:06.941 { 00:10:06.941 "name": "BaseBdev4", 00:10:06.941 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:06.941 "is_configured": true, 00:10:06.941 "data_offset": 2048, 00:10:06.941 "data_size": 63488 00:10:06.941 } 00:10:06.941 ] 00:10:06.941 }' 00:10:06.941 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:06.941 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.510 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.510 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.510 16:48:28 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.510 16:48:28 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:07.510 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.510 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:07.510 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.510 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 28b77698-015a-4499-aa2e-1e0dd87baa52 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.511 [2024-09-29 16:48:29.072550] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:07.511 [2024-09-29 16:48:29.072741] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:07.511 [2024-09-29 16:48:29.072754] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:07.511 NewBaseBdev 00:10:07.511 [2024-09-29 16:48:29.073051] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:07.511 [2024-09-29 16:48:29.073168] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:07.511 [2024-09-29 16:48:29.073179] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:07.511 [2024-09-29 16:48:29.073273] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.511 [ 00:10:07.511 { 00:10:07.511 "name": "NewBaseBdev", 00:10:07.511 "aliases": [ 00:10:07.511 "28b77698-015a-4499-aa2e-1e0dd87baa52" 00:10:07.511 ], 00:10:07.511 "product_name": "Malloc disk", 00:10:07.511 "block_size": 512, 00:10:07.511 "num_blocks": 65536, 00:10:07.511 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:07.511 "assigned_rate_limits": { 00:10:07.511 "rw_ios_per_sec": 0, 00:10:07.511 "rw_mbytes_per_sec": 0, 00:10:07.511 "r_mbytes_per_sec": 0, 00:10:07.511 "w_mbytes_per_sec": 0 00:10:07.511 }, 00:10:07.511 "claimed": true, 00:10:07.511 "claim_type": "exclusive_write", 00:10:07.511 "zoned": false, 00:10:07.511 "supported_io_types": { 00:10:07.511 "read": true, 00:10:07.511 "write": true, 00:10:07.511 "unmap": true, 00:10:07.511 "flush": true, 00:10:07.511 "reset": true, 00:10:07.511 "nvme_admin": false, 00:10:07.511 "nvme_io": false, 00:10:07.511 "nvme_io_md": false, 00:10:07.511 "write_zeroes": true, 00:10:07.511 "zcopy": true, 00:10:07.511 "get_zone_info": false, 00:10:07.511 "zone_management": false, 00:10:07.511 "zone_append": false, 00:10:07.511 "compare": false, 00:10:07.511 "compare_and_write": false, 00:10:07.511 "abort": true, 00:10:07.511 "seek_hole": false, 00:10:07.511 "seek_data": false, 00:10:07.511 "copy": true, 00:10:07.511 "nvme_iov_md": false 00:10:07.511 }, 00:10:07.511 "memory_domains": [ 00:10:07.511 { 00:10:07.511 "dma_device_id": "system", 00:10:07.511 "dma_device_type": 1 00:10:07.511 }, 00:10:07.511 { 00:10:07.511 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:07.511 "dma_device_type": 2 00:10:07.511 } 00:10:07.511 ], 00:10:07.511 "driver_specific": {} 00:10:07.511 } 00:10:07.511 ] 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.511 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.511 "name": "Existed_Raid", 00:10:07.511 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:07.511 "strip_size_kb": 64, 00:10:07.511 "state": "online", 00:10:07.511 "raid_level": "concat", 00:10:07.511 "superblock": true, 00:10:07.511 "num_base_bdevs": 4, 00:10:07.511 "num_base_bdevs_discovered": 4, 00:10:07.511 "num_base_bdevs_operational": 4, 00:10:07.511 "base_bdevs_list": [ 00:10:07.511 { 00:10:07.511 "name": "NewBaseBdev", 00:10:07.511 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:07.511 "is_configured": true, 00:10:07.511 "data_offset": 2048, 00:10:07.511 "data_size": 63488 00:10:07.511 }, 00:10:07.511 { 00:10:07.511 "name": "BaseBdev2", 00:10:07.511 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:07.511 "is_configured": true, 00:10:07.511 "data_offset": 2048, 00:10:07.511 "data_size": 63488 00:10:07.511 }, 00:10:07.511 { 00:10:07.511 "name": "BaseBdev3", 00:10:07.511 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:07.512 "is_configured": true, 00:10:07.512 "data_offset": 2048, 00:10:07.512 "data_size": 63488 00:10:07.512 }, 00:10:07.512 { 00:10:07.512 "name": "BaseBdev4", 00:10:07.512 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:07.512 "is_configured": true, 00:10:07.512 "data_offset": 2048, 00:10:07.512 "data_size": 63488 00:10:07.512 } 00:10:07.512 ] 00:10:07.512 }' 00:10:07.512 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.512 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:08.082 [2024-09-29 16:48:29.544094] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.082 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:08.082 "name": "Existed_Raid", 00:10:08.082 "aliases": [ 00:10:08.082 "1f78b816-1a09-4a63-bc52-7677174013b9" 00:10:08.082 ], 00:10:08.082 "product_name": "Raid Volume", 00:10:08.082 "block_size": 512, 00:10:08.082 "num_blocks": 253952, 00:10:08.082 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:08.082 "assigned_rate_limits": { 00:10:08.082 "rw_ios_per_sec": 0, 00:10:08.082 "rw_mbytes_per_sec": 0, 00:10:08.082 "r_mbytes_per_sec": 0, 00:10:08.082 "w_mbytes_per_sec": 0 00:10:08.082 }, 00:10:08.082 "claimed": false, 00:10:08.082 "zoned": false, 00:10:08.082 "supported_io_types": { 00:10:08.082 "read": true, 00:10:08.082 "write": true, 00:10:08.082 "unmap": true, 00:10:08.082 "flush": true, 00:10:08.082 "reset": true, 00:10:08.082 "nvme_admin": false, 00:10:08.082 "nvme_io": false, 00:10:08.082 "nvme_io_md": false, 00:10:08.082 "write_zeroes": true, 00:10:08.082 "zcopy": false, 00:10:08.082 "get_zone_info": false, 00:10:08.082 "zone_management": false, 00:10:08.082 "zone_append": false, 00:10:08.082 "compare": false, 00:10:08.082 "compare_and_write": false, 00:10:08.082 "abort": false, 00:10:08.082 "seek_hole": false, 00:10:08.082 "seek_data": false, 00:10:08.082 "copy": false, 00:10:08.082 "nvme_iov_md": false 00:10:08.082 }, 00:10:08.082 "memory_domains": [ 00:10:08.082 { 00:10:08.082 "dma_device_id": "system", 00:10:08.082 "dma_device_type": 1 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.082 "dma_device_type": 2 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "dma_device_id": "system", 00:10:08.082 "dma_device_type": 1 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.082 "dma_device_type": 2 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "dma_device_id": "system", 00:10:08.082 "dma_device_type": 1 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.082 "dma_device_type": 2 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "dma_device_id": "system", 00:10:08.082 "dma_device_type": 1 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.082 "dma_device_type": 2 00:10:08.082 } 00:10:08.082 ], 00:10:08.082 "driver_specific": { 00:10:08.082 "raid": { 00:10:08.082 "uuid": "1f78b816-1a09-4a63-bc52-7677174013b9", 00:10:08.082 "strip_size_kb": 64, 00:10:08.082 "state": "online", 00:10:08.082 "raid_level": "concat", 00:10:08.082 "superblock": true, 00:10:08.082 "num_base_bdevs": 4, 00:10:08.082 "num_base_bdevs_discovered": 4, 00:10:08.082 "num_base_bdevs_operational": 4, 00:10:08.082 "base_bdevs_list": [ 00:10:08.082 { 00:10:08.082 "name": "NewBaseBdev", 00:10:08.082 "uuid": "28b77698-015a-4499-aa2e-1e0dd87baa52", 00:10:08.082 "is_configured": true, 00:10:08.082 "data_offset": 2048, 00:10:08.082 "data_size": 63488 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "name": "BaseBdev2", 00:10:08.082 "uuid": "fbef2f67-b1ba-4eb6-b4ef-507b04cdabb0", 00:10:08.082 "is_configured": true, 00:10:08.082 "data_offset": 2048, 00:10:08.082 "data_size": 63488 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "name": "BaseBdev3", 00:10:08.082 "uuid": "900d49d9-ac67-498d-b70d-347289c28a9c", 00:10:08.082 "is_configured": true, 00:10:08.082 "data_offset": 2048, 00:10:08.082 "data_size": 63488 00:10:08.082 }, 00:10:08.082 { 00:10:08.082 "name": "BaseBdev4", 00:10:08.082 "uuid": "c9f32a95-2c1e-4bbb-a8fa-a4d9c48ca323", 00:10:08.082 "is_configured": true, 00:10:08.083 "data_offset": 2048, 00:10:08.083 "data_size": 63488 00:10:08.083 } 00:10:08.083 ] 00:10:08.083 } 00:10:08.083 } 00:10:08.083 }' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:08.083 BaseBdev2 00:10:08.083 BaseBdev3 00:10:08.083 BaseBdev4' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.083 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.343 [2024-09-29 16:48:29.835300] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:08.343 [2024-09-29 16:48:29.835328] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:08.343 [2024-09-29 16:48:29.835402] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:08.343 [2024-09-29 16:48:29.835466] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:08.343 [2024-09-29 16:48:29.835476] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82503 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 82503 ']' 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 82503 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82503 00:10:08.343 killing process with pid 82503 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82503' 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 82503 00:10:08.343 [2024-09-29 16:48:29.873317] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:08.343 16:48:29 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 82503 00:10:08.343 [2024-09-29 16:48:29.912839] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:08.603 16:48:30 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:08.603 00:10:08.603 real 0m9.280s 00:10:08.603 user 0m15.855s 00:10:08.603 sys 0m1.851s 00:10:08.603 16:48:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:08.603 ************************************ 00:10:08.603 END TEST raid_state_function_test_sb 00:10:08.603 ************************************ 00:10:08.603 16:48:30 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.603 16:48:30 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:10:08.603 16:48:30 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:08.603 16:48:30 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:08.603 16:48:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:08.603 ************************************ 00:10:08.603 START TEST raid_superblock_test 00:10:08.603 ************************************ 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 4 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=83151 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 83151 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 83151 ']' 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:08.603 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:08.603 16:48:30 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:08.863 [2024-09-29 16:48:30.313892] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:08.863 [2024-09-29 16:48:30.314511] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83151 ] 00:10:08.863 [2024-09-29 16:48:30.438437] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:08.863 [2024-09-29 16:48:30.482957] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:08.863 [2024-09-29 16:48:30.524819] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:08.863 [2024-09-29 16:48:30.524938] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.802 malloc1 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.802 [2024-09-29 16:48:31.150950] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:09.802 [2024-09-29 16:48:31.151053] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:09.802 [2024-09-29 16:48:31.151090] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:09.802 [2024-09-29 16:48:31.151142] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:09.802 [2024-09-29 16:48:31.153170] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:09.802 [2024-09-29 16:48:31.153256] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:09.802 pt1 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.802 malloc2 00:10:09.802 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.803 [2024-09-29 16:48:31.197070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:09.803 [2024-09-29 16:48:31.197154] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:09.803 [2024-09-29 16:48:31.197183] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:09.803 [2024-09-29 16:48:31.197203] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:09.803 [2024-09-29 16:48:31.200958] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:09.803 [2024-09-29 16:48:31.201053] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:09.803 pt2 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.803 malloc3 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.803 [2024-09-29 16:48:31.225859] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:09.803 [2024-09-29 16:48:31.225960] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:09.803 [2024-09-29 16:48:31.225994] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:09.803 [2024-09-29 16:48:31.226022] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:09.803 [2024-09-29 16:48:31.228056] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:09.803 [2024-09-29 16:48:31.228145] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:09.803 pt3 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.803 malloc4 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.803 [2024-09-29 16:48:31.258263] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:09.803 [2024-09-29 16:48:31.258363] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:09.803 [2024-09-29 16:48:31.258393] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:09.803 [2024-09-29 16:48:31.258424] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:09.803 [2024-09-29 16:48:31.260512] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:09.803 [2024-09-29 16:48:31.260583] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:09.803 pt4 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.803 [2024-09-29 16:48:31.270289] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:09.803 [2024-09-29 16:48:31.272158] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:09.803 [2024-09-29 16:48:31.272285] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:09.803 [2024-09-29 16:48:31.272363] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:09.803 [2024-09-29 16:48:31.272558] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:09.803 [2024-09-29 16:48:31.272608] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:09.803 [2024-09-29 16:48:31.272891] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:09.803 [2024-09-29 16:48:31.273071] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:09.803 [2024-09-29 16:48:31.273113] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:09.803 [2024-09-29 16:48:31.273303] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.803 "name": "raid_bdev1", 00:10:09.803 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:09.803 "strip_size_kb": 64, 00:10:09.803 "state": "online", 00:10:09.803 "raid_level": "concat", 00:10:09.803 "superblock": true, 00:10:09.803 "num_base_bdevs": 4, 00:10:09.803 "num_base_bdevs_discovered": 4, 00:10:09.803 "num_base_bdevs_operational": 4, 00:10:09.803 "base_bdevs_list": [ 00:10:09.803 { 00:10:09.803 "name": "pt1", 00:10:09.803 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:09.803 "is_configured": true, 00:10:09.803 "data_offset": 2048, 00:10:09.803 "data_size": 63488 00:10:09.803 }, 00:10:09.803 { 00:10:09.803 "name": "pt2", 00:10:09.803 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:09.803 "is_configured": true, 00:10:09.803 "data_offset": 2048, 00:10:09.803 "data_size": 63488 00:10:09.803 }, 00:10:09.803 { 00:10:09.803 "name": "pt3", 00:10:09.803 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:09.803 "is_configured": true, 00:10:09.803 "data_offset": 2048, 00:10:09.803 "data_size": 63488 00:10:09.803 }, 00:10:09.803 { 00:10:09.803 "name": "pt4", 00:10:09.803 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:09.803 "is_configured": true, 00:10:09.803 "data_offset": 2048, 00:10:09.803 "data_size": 63488 00:10:09.803 } 00:10:09.803 ] 00:10:09.803 }' 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.803 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.063 [2024-09-29 16:48:31.633939] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.063 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:10.063 "name": "raid_bdev1", 00:10:10.063 "aliases": [ 00:10:10.063 "159eedc5-bcf3-4a28-a646-6efa1c45f1ad" 00:10:10.063 ], 00:10:10.063 "product_name": "Raid Volume", 00:10:10.063 "block_size": 512, 00:10:10.063 "num_blocks": 253952, 00:10:10.063 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:10.063 "assigned_rate_limits": { 00:10:10.063 "rw_ios_per_sec": 0, 00:10:10.064 "rw_mbytes_per_sec": 0, 00:10:10.064 "r_mbytes_per_sec": 0, 00:10:10.064 "w_mbytes_per_sec": 0 00:10:10.064 }, 00:10:10.064 "claimed": false, 00:10:10.064 "zoned": false, 00:10:10.064 "supported_io_types": { 00:10:10.064 "read": true, 00:10:10.064 "write": true, 00:10:10.064 "unmap": true, 00:10:10.064 "flush": true, 00:10:10.064 "reset": true, 00:10:10.064 "nvme_admin": false, 00:10:10.064 "nvme_io": false, 00:10:10.064 "nvme_io_md": false, 00:10:10.064 "write_zeroes": true, 00:10:10.064 "zcopy": false, 00:10:10.064 "get_zone_info": false, 00:10:10.064 "zone_management": false, 00:10:10.064 "zone_append": false, 00:10:10.064 "compare": false, 00:10:10.064 "compare_and_write": false, 00:10:10.064 "abort": false, 00:10:10.064 "seek_hole": false, 00:10:10.064 "seek_data": false, 00:10:10.064 "copy": false, 00:10:10.064 "nvme_iov_md": false 00:10:10.064 }, 00:10:10.064 "memory_domains": [ 00:10:10.064 { 00:10:10.064 "dma_device_id": "system", 00:10:10.064 "dma_device_type": 1 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.064 "dma_device_type": 2 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "dma_device_id": "system", 00:10:10.064 "dma_device_type": 1 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.064 "dma_device_type": 2 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "dma_device_id": "system", 00:10:10.064 "dma_device_type": 1 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.064 "dma_device_type": 2 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "dma_device_id": "system", 00:10:10.064 "dma_device_type": 1 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.064 "dma_device_type": 2 00:10:10.064 } 00:10:10.064 ], 00:10:10.064 "driver_specific": { 00:10:10.064 "raid": { 00:10:10.064 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:10.064 "strip_size_kb": 64, 00:10:10.064 "state": "online", 00:10:10.064 "raid_level": "concat", 00:10:10.064 "superblock": true, 00:10:10.064 "num_base_bdevs": 4, 00:10:10.064 "num_base_bdevs_discovered": 4, 00:10:10.064 "num_base_bdevs_operational": 4, 00:10:10.064 "base_bdevs_list": [ 00:10:10.064 { 00:10:10.064 "name": "pt1", 00:10:10.064 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:10.064 "is_configured": true, 00:10:10.064 "data_offset": 2048, 00:10:10.064 "data_size": 63488 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "name": "pt2", 00:10:10.064 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:10.064 "is_configured": true, 00:10:10.064 "data_offset": 2048, 00:10:10.064 "data_size": 63488 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "name": "pt3", 00:10:10.064 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:10.064 "is_configured": true, 00:10:10.064 "data_offset": 2048, 00:10:10.064 "data_size": 63488 00:10:10.064 }, 00:10:10.064 { 00:10:10.064 "name": "pt4", 00:10:10.064 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:10.064 "is_configured": true, 00:10:10.064 "data_offset": 2048, 00:10:10.064 "data_size": 63488 00:10:10.064 } 00:10:10.064 ] 00:10:10.064 } 00:10:10.064 } 00:10:10.064 }' 00:10:10.064 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:10.064 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:10.064 pt2 00:10:10.064 pt3 00:10:10.064 pt4' 00:10:10.064 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:10.324 [2024-09-29 16:48:31.945344] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=159eedc5-bcf3-4a28-a646-6efa1c45f1ad 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 159eedc5-bcf3-4a28-a646-6efa1c45f1ad ']' 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.324 [2024-09-29 16:48:31.989008] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:10.324 [2024-09-29 16:48:31.989076] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:10.324 [2024-09-29 16:48:31.989172] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:10.324 [2024-09-29 16:48:31.989275] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:10.324 [2024-09-29 16:48:31.989340] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:10.324 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.585 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.585 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:31 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 16:48:31 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 [2024-09-29 16:48:32.132791] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:10.585 [2024-09-29 16:48:32.134565] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:10.585 [2024-09-29 16:48:32.134606] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:10.585 [2024-09-29 16:48:32.134640] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:10.585 [2024-09-29 16:48:32.134688] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:10.585 [2024-09-29 16:48:32.134742] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:10.585 [2024-09-29 16:48:32.134762] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:10.585 [2024-09-29 16:48:32.134778] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:10.585 [2024-09-29 16:48:32.134791] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:10.585 [2024-09-29 16:48:32.134800] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:10.585 request: 00:10:10.585 { 00:10:10.585 "name": "raid_bdev1", 00:10:10.585 "raid_level": "concat", 00:10:10.585 "base_bdevs": [ 00:10:10.585 "malloc1", 00:10:10.585 "malloc2", 00:10:10.585 "malloc3", 00:10:10.585 "malloc4" 00:10:10.585 ], 00:10:10.585 "strip_size_kb": 64, 00:10:10.585 "superblock": false, 00:10:10.585 "method": "bdev_raid_create", 00:10:10.585 "req_id": 1 00:10:10.585 } 00:10:10.585 Got JSON-RPC error response 00:10:10.585 response: 00:10:10.585 { 00:10:10.585 "code": -17, 00:10:10.585 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:10.585 } 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.585 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.586 [2024-09-29 16:48:32.200614] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:10.586 [2024-09-29 16:48:32.200698] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:10.586 [2024-09-29 16:48:32.200748] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:10.586 [2024-09-29 16:48:32.200777] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:10.586 [2024-09-29 16:48:32.202869] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:10.586 [2024-09-29 16:48:32.202946] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:10.586 [2024-09-29 16:48:32.203033] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:10.586 [2024-09-29 16:48:32.203089] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:10.586 pt1 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.586 "name": "raid_bdev1", 00:10:10.586 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:10.586 "strip_size_kb": 64, 00:10:10.586 "state": "configuring", 00:10:10.586 "raid_level": "concat", 00:10:10.586 "superblock": true, 00:10:10.586 "num_base_bdevs": 4, 00:10:10.586 "num_base_bdevs_discovered": 1, 00:10:10.586 "num_base_bdevs_operational": 4, 00:10:10.586 "base_bdevs_list": [ 00:10:10.586 { 00:10:10.586 "name": "pt1", 00:10:10.586 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:10.586 "is_configured": true, 00:10:10.586 "data_offset": 2048, 00:10:10.586 "data_size": 63488 00:10:10.586 }, 00:10:10.586 { 00:10:10.586 "name": null, 00:10:10.586 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:10.586 "is_configured": false, 00:10:10.586 "data_offset": 2048, 00:10:10.586 "data_size": 63488 00:10:10.586 }, 00:10:10.586 { 00:10:10.586 "name": null, 00:10:10.586 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:10.586 "is_configured": false, 00:10:10.586 "data_offset": 2048, 00:10:10.586 "data_size": 63488 00:10:10.586 }, 00:10:10.586 { 00:10:10.586 "name": null, 00:10:10.586 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:10.586 "is_configured": false, 00:10:10.586 "data_offset": 2048, 00:10:10.586 "data_size": 63488 00:10:10.586 } 00:10:10.586 ] 00:10:10.586 }' 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.586 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.155 [2024-09-29 16:48:32.571966] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:11.155 [2024-09-29 16:48:32.572054] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.155 [2024-09-29 16:48:32.572089] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:11.155 [2024-09-29 16:48:32.572116] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.155 [2024-09-29 16:48:32.572516] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.155 [2024-09-29 16:48:32.572547] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:11.155 [2024-09-29 16:48:32.572613] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:11.155 [2024-09-29 16:48:32.572640] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:11.155 pt2 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.155 [2024-09-29 16:48:32.579986] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.155 "name": "raid_bdev1", 00:10:11.155 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:11.155 "strip_size_kb": 64, 00:10:11.155 "state": "configuring", 00:10:11.155 "raid_level": "concat", 00:10:11.155 "superblock": true, 00:10:11.155 "num_base_bdevs": 4, 00:10:11.155 "num_base_bdevs_discovered": 1, 00:10:11.155 "num_base_bdevs_operational": 4, 00:10:11.155 "base_bdevs_list": [ 00:10:11.155 { 00:10:11.155 "name": "pt1", 00:10:11.155 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:11.155 "is_configured": true, 00:10:11.155 "data_offset": 2048, 00:10:11.155 "data_size": 63488 00:10:11.155 }, 00:10:11.155 { 00:10:11.155 "name": null, 00:10:11.155 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.155 "is_configured": false, 00:10:11.155 "data_offset": 0, 00:10:11.155 "data_size": 63488 00:10:11.155 }, 00:10:11.155 { 00:10:11.155 "name": null, 00:10:11.155 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.155 "is_configured": false, 00:10:11.155 "data_offset": 2048, 00:10:11.155 "data_size": 63488 00:10:11.155 }, 00:10:11.155 { 00:10:11.155 "name": null, 00:10:11.155 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.155 "is_configured": false, 00:10:11.155 "data_offset": 2048, 00:10:11.155 "data_size": 63488 00:10:11.155 } 00:10:11.155 ] 00:10:11.155 }' 00:10:11.155 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.156 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.416 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:11.416 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:11.416 16:48:32 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:11.416 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.416 16:48:32 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.416 [2024-09-29 16:48:32.999271] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:11.416 [2024-09-29 16:48:32.999369] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.416 [2024-09-29 16:48:32.999389] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:11.416 [2024-09-29 16:48:32.999399] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.416 [2024-09-29 16:48:32.999797] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.416 [2024-09-29 16:48:32.999820] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:11.416 [2024-09-29 16:48:32.999884] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:11.416 [2024-09-29 16:48:32.999905] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:11.416 pt2 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.416 [2024-09-29 16:48:33.011214] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:11.416 [2024-09-29 16:48:33.011264] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.416 [2024-09-29 16:48:33.011280] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:11.416 [2024-09-29 16:48:33.011299] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.416 [2024-09-29 16:48:33.011590] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.416 [2024-09-29 16:48:33.011607] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:11.416 [2024-09-29 16:48:33.011660] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:11.416 [2024-09-29 16:48:33.011683] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:11.416 pt3 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.416 [2024-09-29 16:48:33.023216] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:11.416 [2024-09-29 16:48:33.023262] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:11.416 [2024-09-29 16:48:33.023276] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:11.416 [2024-09-29 16:48:33.023285] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:11.416 [2024-09-29 16:48:33.023555] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:11.416 [2024-09-29 16:48:33.023573] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:11.416 [2024-09-29 16:48:33.023618] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:11.416 [2024-09-29 16:48:33.023635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:11.416 [2024-09-29 16:48:33.023743] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:11.416 [2024-09-29 16:48:33.023755] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:11.416 [2024-09-29 16:48:33.023975] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:11.416 [2024-09-29 16:48:33.024103] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:11.416 [2024-09-29 16:48:33.024111] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:11.416 [2024-09-29 16:48:33.024207] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:11.416 pt4 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.416 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.416 "name": "raid_bdev1", 00:10:11.416 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:11.416 "strip_size_kb": 64, 00:10:11.416 "state": "online", 00:10:11.416 "raid_level": "concat", 00:10:11.416 "superblock": true, 00:10:11.416 "num_base_bdevs": 4, 00:10:11.416 "num_base_bdevs_discovered": 4, 00:10:11.416 "num_base_bdevs_operational": 4, 00:10:11.416 "base_bdevs_list": [ 00:10:11.416 { 00:10:11.416 "name": "pt1", 00:10:11.416 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:11.416 "is_configured": true, 00:10:11.416 "data_offset": 2048, 00:10:11.416 "data_size": 63488 00:10:11.416 }, 00:10:11.416 { 00:10:11.416 "name": "pt2", 00:10:11.416 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.416 "is_configured": true, 00:10:11.416 "data_offset": 2048, 00:10:11.417 "data_size": 63488 00:10:11.417 }, 00:10:11.417 { 00:10:11.417 "name": "pt3", 00:10:11.417 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.417 "is_configured": true, 00:10:11.417 "data_offset": 2048, 00:10:11.417 "data_size": 63488 00:10:11.417 }, 00:10:11.417 { 00:10:11.417 "name": "pt4", 00:10:11.417 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.417 "is_configured": true, 00:10:11.417 "data_offset": 2048, 00:10:11.417 "data_size": 63488 00:10:11.417 } 00:10:11.417 ] 00:10:11.417 }' 00:10:11.417 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.417 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.985 [2024-09-29 16:48:33.446852] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.985 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:11.985 "name": "raid_bdev1", 00:10:11.985 "aliases": [ 00:10:11.985 "159eedc5-bcf3-4a28-a646-6efa1c45f1ad" 00:10:11.985 ], 00:10:11.985 "product_name": "Raid Volume", 00:10:11.985 "block_size": 512, 00:10:11.985 "num_blocks": 253952, 00:10:11.985 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:11.985 "assigned_rate_limits": { 00:10:11.985 "rw_ios_per_sec": 0, 00:10:11.985 "rw_mbytes_per_sec": 0, 00:10:11.986 "r_mbytes_per_sec": 0, 00:10:11.986 "w_mbytes_per_sec": 0 00:10:11.986 }, 00:10:11.986 "claimed": false, 00:10:11.986 "zoned": false, 00:10:11.986 "supported_io_types": { 00:10:11.986 "read": true, 00:10:11.986 "write": true, 00:10:11.986 "unmap": true, 00:10:11.986 "flush": true, 00:10:11.986 "reset": true, 00:10:11.986 "nvme_admin": false, 00:10:11.986 "nvme_io": false, 00:10:11.986 "nvme_io_md": false, 00:10:11.986 "write_zeroes": true, 00:10:11.986 "zcopy": false, 00:10:11.986 "get_zone_info": false, 00:10:11.986 "zone_management": false, 00:10:11.986 "zone_append": false, 00:10:11.986 "compare": false, 00:10:11.986 "compare_and_write": false, 00:10:11.986 "abort": false, 00:10:11.986 "seek_hole": false, 00:10:11.986 "seek_data": false, 00:10:11.986 "copy": false, 00:10:11.986 "nvme_iov_md": false 00:10:11.986 }, 00:10:11.986 "memory_domains": [ 00:10:11.986 { 00:10:11.986 "dma_device_id": "system", 00:10:11.986 "dma_device_type": 1 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.986 "dma_device_type": 2 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "dma_device_id": "system", 00:10:11.986 "dma_device_type": 1 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.986 "dma_device_type": 2 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "dma_device_id": "system", 00:10:11.986 "dma_device_type": 1 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.986 "dma_device_type": 2 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "dma_device_id": "system", 00:10:11.986 "dma_device_type": 1 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.986 "dma_device_type": 2 00:10:11.986 } 00:10:11.986 ], 00:10:11.986 "driver_specific": { 00:10:11.986 "raid": { 00:10:11.986 "uuid": "159eedc5-bcf3-4a28-a646-6efa1c45f1ad", 00:10:11.986 "strip_size_kb": 64, 00:10:11.986 "state": "online", 00:10:11.986 "raid_level": "concat", 00:10:11.986 "superblock": true, 00:10:11.986 "num_base_bdevs": 4, 00:10:11.986 "num_base_bdevs_discovered": 4, 00:10:11.986 "num_base_bdevs_operational": 4, 00:10:11.986 "base_bdevs_list": [ 00:10:11.986 { 00:10:11.986 "name": "pt1", 00:10:11.986 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:11.986 "is_configured": true, 00:10:11.986 "data_offset": 2048, 00:10:11.986 "data_size": 63488 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "name": "pt2", 00:10:11.986 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:11.986 "is_configured": true, 00:10:11.986 "data_offset": 2048, 00:10:11.986 "data_size": 63488 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "name": "pt3", 00:10:11.986 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:11.986 "is_configured": true, 00:10:11.986 "data_offset": 2048, 00:10:11.986 "data_size": 63488 00:10:11.986 }, 00:10:11.986 { 00:10:11.986 "name": "pt4", 00:10:11.986 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:11.986 "is_configured": true, 00:10:11.986 "data_offset": 2048, 00:10:11.986 "data_size": 63488 00:10:11.986 } 00:10:11.986 ] 00:10:11.986 } 00:10:11.986 } 00:10:11.986 }' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:11.986 pt2 00:10:11.986 pt3 00:10:11.986 pt4' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.986 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:12.246 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:12.247 [2024-09-29 16:48:33.730320] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 159eedc5-bcf3-4a28-a646-6efa1c45f1ad '!=' 159eedc5-bcf3-4a28-a646-6efa1c45f1ad ']' 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 83151 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 83151 ']' 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 83151 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83151 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83151' 00:10:12.247 killing process with pid 83151 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 83151 00:10:12.247 [2024-09-29 16:48:33.818361] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:12.247 [2024-09-29 16:48:33.818492] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:12.247 16:48:33 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 83151 00:10:12.247 [2024-09-29 16:48:33.818588] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:12.247 [2024-09-29 16:48:33.818603] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:12.247 [2024-09-29 16:48:33.860703] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:12.507 16:48:34 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:12.507 00:10:12.507 real 0m3.874s 00:10:12.507 user 0m6.051s 00:10:12.507 sys 0m0.825s 00:10:12.507 16:48:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:12.507 16:48:34 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.507 ************************************ 00:10:12.507 END TEST raid_superblock_test 00:10:12.507 ************************************ 00:10:12.507 16:48:34 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:10:12.507 16:48:34 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:12.507 16:48:34 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:12.507 16:48:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:12.507 ************************************ 00:10:12.507 START TEST raid_read_error_test 00:10:12.507 ************************************ 00:10:12.507 16:48:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 read 00:10:12.507 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:12.507 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:12.507 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:12.507 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.6pXAGbsgH1 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83398 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83398 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 83398 ']' 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:12.767 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:12.767 16:48:34 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:12.767 [2024-09-29 16:48:34.268193] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:12.767 [2024-09-29 16:48:34.268683] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83398 ] 00:10:12.767 [2024-09-29 16:48:34.394410] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:13.027 [2024-09-29 16:48:34.439250] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:13.027 [2024-09-29 16:48:34.481173] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:13.027 [2024-09-29 16:48:34.481206] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.596 BaseBdev1_malloc 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.596 true 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.596 [2024-09-29 16:48:35.130992] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:13.596 [2024-09-29 16:48:35.131051] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.596 [2024-09-29 16:48:35.131071] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:13.596 [2024-09-29 16:48:35.131080] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.596 [2024-09-29 16:48:35.133170] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.596 [2024-09-29 16:48:35.133217] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:13.596 BaseBdev1 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.596 BaseBdev2_malloc 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.596 true 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.596 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.596 [2024-09-29 16:48:35.179131] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:13.596 [2024-09-29 16:48:35.179244] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.597 [2024-09-29 16:48:35.179268] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:13.597 [2024-09-29 16:48:35.179277] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.597 [2024-09-29 16:48:35.181358] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.597 [2024-09-29 16:48:35.181395] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:13.597 BaseBdev2 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.597 BaseBdev3_malloc 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.597 true 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.597 [2024-09-29 16:48:35.219900] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:13.597 [2024-09-29 16:48:35.219945] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.597 [2024-09-29 16:48:35.219964] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:13.597 [2024-09-29 16:48:35.219972] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.597 [2024-09-29 16:48:35.222061] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.597 [2024-09-29 16:48:35.222096] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:13.597 BaseBdev3 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.597 BaseBdev4_malloc 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.597 true 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.597 [2024-09-29 16:48:35.260472] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:13.597 [2024-09-29 16:48:35.260569] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:13.597 [2024-09-29 16:48:35.260593] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:13.597 [2024-09-29 16:48:35.260602] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:13.597 [2024-09-29 16:48:35.262617] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:13.597 [2024-09-29 16:48:35.262654] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:13.597 BaseBdev4 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.597 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.857 [2024-09-29 16:48:35.272511] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:13.857 [2024-09-29 16:48:35.274405] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:13.857 [2024-09-29 16:48:35.274477] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:13.857 [2024-09-29 16:48:35.274545] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:13.857 [2024-09-29 16:48:35.274736] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:13.857 [2024-09-29 16:48:35.274764] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:13.857 [2024-09-29 16:48:35.275047] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:13.857 [2024-09-29 16:48:35.275190] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:13.857 [2024-09-29 16:48:35.275209] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:13.857 [2024-09-29 16:48:35.275348] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.857 "name": "raid_bdev1", 00:10:13.857 "uuid": "471d95ae-f1a2-4023-874f-f044147e9c59", 00:10:13.857 "strip_size_kb": 64, 00:10:13.857 "state": "online", 00:10:13.857 "raid_level": "concat", 00:10:13.857 "superblock": true, 00:10:13.857 "num_base_bdevs": 4, 00:10:13.857 "num_base_bdevs_discovered": 4, 00:10:13.857 "num_base_bdevs_operational": 4, 00:10:13.857 "base_bdevs_list": [ 00:10:13.857 { 00:10:13.857 "name": "BaseBdev1", 00:10:13.857 "uuid": "5cbe8b96-ccae-570e-a434-cfd5d25ee7a6", 00:10:13.857 "is_configured": true, 00:10:13.857 "data_offset": 2048, 00:10:13.857 "data_size": 63488 00:10:13.857 }, 00:10:13.857 { 00:10:13.857 "name": "BaseBdev2", 00:10:13.857 "uuid": "5fecef23-4ad8-5442-b872-1ae5b79d548f", 00:10:13.857 "is_configured": true, 00:10:13.857 "data_offset": 2048, 00:10:13.857 "data_size": 63488 00:10:13.857 }, 00:10:13.857 { 00:10:13.857 "name": "BaseBdev3", 00:10:13.857 "uuid": "03e46628-ecca-5b65-9b60-b28590fc37eb", 00:10:13.857 "is_configured": true, 00:10:13.857 "data_offset": 2048, 00:10:13.857 "data_size": 63488 00:10:13.857 }, 00:10:13.857 { 00:10:13.857 "name": "BaseBdev4", 00:10:13.857 "uuid": "b1bbf02b-4121-56bf-bf9c-58a0a35c9a47", 00:10:13.857 "is_configured": true, 00:10:13.857 "data_offset": 2048, 00:10:13.857 "data_size": 63488 00:10:13.857 } 00:10:13.857 ] 00:10:13.857 }' 00:10:13.857 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.858 16:48:35 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:14.117 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:14.117 16:48:35 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:14.117 [2024-09-29 16:48:35.760058] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.077 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:15.078 16:48:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.078 16:48:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.078 16:48:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.353 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.353 "name": "raid_bdev1", 00:10:15.353 "uuid": "471d95ae-f1a2-4023-874f-f044147e9c59", 00:10:15.353 "strip_size_kb": 64, 00:10:15.353 "state": "online", 00:10:15.353 "raid_level": "concat", 00:10:15.353 "superblock": true, 00:10:15.354 "num_base_bdevs": 4, 00:10:15.354 "num_base_bdevs_discovered": 4, 00:10:15.354 "num_base_bdevs_operational": 4, 00:10:15.354 "base_bdevs_list": [ 00:10:15.354 { 00:10:15.354 "name": "BaseBdev1", 00:10:15.354 "uuid": "5cbe8b96-ccae-570e-a434-cfd5d25ee7a6", 00:10:15.354 "is_configured": true, 00:10:15.354 "data_offset": 2048, 00:10:15.354 "data_size": 63488 00:10:15.354 }, 00:10:15.354 { 00:10:15.354 "name": "BaseBdev2", 00:10:15.354 "uuid": "5fecef23-4ad8-5442-b872-1ae5b79d548f", 00:10:15.354 "is_configured": true, 00:10:15.354 "data_offset": 2048, 00:10:15.354 "data_size": 63488 00:10:15.354 }, 00:10:15.354 { 00:10:15.354 "name": "BaseBdev3", 00:10:15.354 "uuid": "03e46628-ecca-5b65-9b60-b28590fc37eb", 00:10:15.354 "is_configured": true, 00:10:15.354 "data_offset": 2048, 00:10:15.354 "data_size": 63488 00:10:15.354 }, 00:10:15.354 { 00:10:15.354 "name": "BaseBdev4", 00:10:15.354 "uuid": "b1bbf02b-4121-56bf-bf9c-58a0a35c9a47", 00:10:15.354 "is_configured": true, 00:10:15.354 "data_offset": 2048, 00:10:15.354 "data_size": 63488 00:10:15.354 } 00:10:15.354 ] 00:10:15.354 }' 00:10:15.354 16:48:36 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.354 16:48:36 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.628 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:15.628 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.628 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.628 [2024-09-29 16:48:37.123554] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:15.628 [2024-09-29 16:48:37.123588] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:15.628 [2024-09-29 16:48:37.126036] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:15.628 [2024-09-29 16:48:37.126086] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:15.628 [2024-09-29 16:48:37.126129] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:15.628 [2024-09-29 16:48:37.126138] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:15.628 { 00:10:15.628 "results": [ 00:10:15.628 { 00:10:15.628 "job": "raid_bdev1", 00:10:15.628 "core_mask": "0x1", 00:10:15.628 "workload": "randrw", 00:10:15.628 "percentage": 50, 00:10:15.628 "status": "finished", 00:10:15.628 "queue_depth": 1, 00:10:15.628 "io_size": 131072, 00:10:15.628 "runtime": 1.364304, 00:10:15.628 "iops": 16975.688702811105, 00:10:15.628 "mibps": 2121.961087851388, 00:10:15.628 "io_failed": 1, 00:10:15.628 "io_timeout": 0, 00:10:15.628 "avg_latency_us": 81.74418968492624, 00:10:15.628 "min_latency_us": 24.705676855895195, 00:10:15.629 "max_latency_us": 1352.216593886463 00:10:15.629 } 00:10:15.629 ], 00:10:15.629 "core_count": 1 00:10:15.629 } 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83398 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 83398 ']' 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 83398 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83398 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83398' 00:10:15.629 killing process with pid 83398 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 83398 00:10:15.629 [2024-09-29 16:48:37.172581] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:15.629 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 83398 00:10:15.629 [2024-09-29 16:48:37.206574] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.6pXAGbsgH1 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:15.889 ************************************ 00:10:15.889 END TEST raid_read_error_test 00:10:15.889 ************************************ 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:10:15.889 00:10:15.889 real 0m3.279s 00:10:15.889 user 0m4.088s 00:10:15.889 sys 0m0.518s 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:15.889 16:48:37 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:15.889 16:48:37 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:10:15.889 16:48:37 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:15.889 16:48:37 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:15.889 16:48:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:15.889 ************************************ 00:10:15.889 START TEST raid_write_error_test 00:10:15.889 ************************************ 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 write 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.wR9y7vROk1 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83528 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83528 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 83528 ']' 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:15.889 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:15.889 16:48:37 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.150 [2024-09-29 16:48:37.624250] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:16.150 [2024-09-29 16:48:37.624397] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83528 ] 00:10:16.150 [2024-09-29 16:48:37.768237] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:16.150 [2024-09-29 16:48:37.814354] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:16.409 [2024-09-29 16:48:37.856460] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:16.409 [2024-09-29 16:48:37.856496] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 BaseBdev1_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 true 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 [2024-09-29 16:48:38.478542] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:16.980 [2024-09-29 16:48:38.478606] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:16.980 [2024-09-29 16:48:38.478635] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:16.980 [2024-09-29 16:48:38.478646] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:16.980 [2024-09-29 16:48:38.480799] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:16.980 [2024-09-29 16:48:38.480834] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:16.980 BaseBdev1 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 BaseBdev2_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 true 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 [2024-09-29 16:48:38.536150] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:16.980 [2024-09-29 16:48:38.536231] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:16.980 [2024-09-29 16:48:38.536262] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:16.980 [2024-09-29 16:48:38.536275] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:16.980 [2024-09-29 16:48:38.539441] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:16.980 [2024-09-29 16:48:38.539490] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:16.980 BaseBdev2 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 BaseBdev3_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 true 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 [2024-09-29 16:48:38.576998] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:16.980 [2024-09-29 16:48:38.577041] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:16.980 [2024-09-29 16:48:38.577074] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:16.980 [2024-09-29 16:48:38.577082] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:16.980 [2024-09-29 16:48:38.579094] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:16.980 [2024-09-29 16:48:38.579182] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:16.980 BaseBdev3 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 BaseBdev4_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 true 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.980 [2024-09-29 16:48:38.617370] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:16.980 [2024-09-29 16:48:38.617413] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:16.980 [2024-09-29 16:48:38.617449] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:16.980 [2024-09-29 16:48:38.617457] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:16.980 [2024-09-29 16:48:38.619501] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:16.980 [2024-09-29 16:48:38.619537] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:16.980 BaseBdev4 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.980 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:16.981 [2024-09-29 16:48:38.629410] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:16.981 [2024-09-29 16:48:38.631272] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:16.981 [2024-09-29 16:48:38.631346] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:16.981 [2024-09-29 16:48:38.631406] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:16.981 [2024-09-29 16:48:38.631596] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:16.981 [2024-09-29 16:48:38.631608] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:16.981 [2024-09-29 16:48:38.631866] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:16.981 [2024-09-29 16:48:38.631993] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:16.981 [2024-09-29 16:48:38.632014] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:16.981 [2024-09-29 16:48:38.632143] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:16.981 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.241 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.241 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.241 "name": "raid_bdev1", 00:10:17.241 "uuid": "fefa6181-a524-433d-be91-16f319c07e3f", 00:10:17.241 "strip_size_kb": 64, 00:10:17.241 "state": "online", 00:10:17.241 "raid_level": "concat", 00:10:17.241 "superblock": true, 00:10:17.241 "num_base_bdevs": 4, 00:10:17.241 "num_base_bdevs_discovered": 4, 00:10:17.241 "num_base_bdevs_operational": 4, 00:10:17.241 "base_bdevs_list": [ 00:10:17.241 { 00:10:17.241 "name": "BaseBdev1", 00:10:17.241 "uuid": "58c18664-8a32-53cf-8c77-60735e00e2d2", 00:10:17.241 "is_configured": true, 00:10:17.241 "data_offset": 2048, 00:10:17.241 "data_size": 63488 00:10:17.241 }, 00:10:17.241 { 00:10:17.241 "name": "BaseBdev2", 00:10:17.241 "uuid": "89c0b877-28f0-5ee2-9d43-cc9e3190f8c6", 00:10:17.241 "is_configured": true, 00:10:17.241 "data_offset": 2048, 00:10:17.241 "data_size": 63488 00:10:17.241 }, 00:10:17.241 { 00:10:17.241 "name": "BaseBdev3", 00:10:17.241 "uuid": "68a792ed-b422-5c62-8ac6-11ba6e9c0b34", 00:10:17.241 "is_configured": true, 00:10:17.241 "data_offset": 2048, 00:10:17.241 "data_size": 63488 00:10:17.241 }, 00:10:17.241 { 00:10:17.241 "name": "BaseBdev4", 00:10:17.241 "uuid": "5a119f58-4847-53a1-8a33-f7f26537f70e", 00:10:17.241 "is_configured": true, 00:10:17.241 "data_offset": 2048, 00:10:17.241 "data_size": 63488 00:10:17.241 } 00:10:17.241 ] 00:10:17.241 }' 00:10:17.241 16:48:38 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.241 16:48:38 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.500 16:48:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:17.500 16:48:39 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:17.500 [2024-09-29 16:48:39.148882] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.440 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.700 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.700 "name": "raid_bdev1", 00:10:18.700 "uuid": "fefa6181-a524-433d-be91-16f319c07e3f", 00:10:18.700 "strip_size_kb": 64, 00:10:18.700 "state": "online", 00:10:18.700 "raid_level": "concat", 00:10:18.700 "superblock": true, 00:10:18.700 "num_base_bdevs": 4, 00:10:18.700 "num_base_bdevs_discovered": 4, 00:10:18.700 "num_base_bdevs_operational": 4, 00:10:18.700 "base_bdevs_list": [ 00:10:18.700 { 00:10:18.700 "name": "BaseBdev1", 00:10:18.700 "uuid": "58c18664-8a32-53cf-8c77-60735e00e2d2", 00:10:18.700 "is_configured": true, 00:10:18.700 "data_offset": 2048, 00:10:18.700 "data_size": 63488 00:10:18.700 }, 00:10:18.700 { 00:10:18.700 "name": "BaseBdev2", 00:10:18.700 "uuid": "89c0b877-28f0-5ee2-9d43-cc9e3190f8c6", 00:10:18.700 "is_configured": true, 00:10:18.700 "data_offset": 2048, 00:10:18.700 "data_size": 63488 00:10:18.700 }, 00:10:18.700 { 00:10:18.700 "name": "BaseBdev3", 00:10:18.700 "uuid": "68a792ed-b422-5c62-8ac6-11ba6e9c0b34", 00:10:18.700 "is_configured": true, 00:10:18.700 "data_offset": 2048, 00:10:18.700 "data_size": 63488 00:10:18.700 }, 00:10:18.700 { 00:10:18.700 "name": "BaseBdev4", 00:10:18.700 "uuid": "5a119f58-4847-53a1-8a33-f7f26537f70e", 00:10:18.700 "is_configured": true, 00:10:18.700 "data_offset": 2048, 00:10:18.700 "data_size": 63488 00:10:18.700 } 00:10:18.700 ] 00:10:18.700 }' 00:10:18.700 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.700 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.960 [2024-09-29 16:48:40.532699] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:18.960 [2024-09-29 16:48:40.532806] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:18.960 [2024-09-29 16:48:40.535296] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:18.960 [2024-09-29 16:48:40.535389] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:18.960 [2024-09-29 16:48:40.535456] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:18.960 [2024-09-29 16:48:40.535509] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:18.960 { 00:10:18.960 "results": [ 00:10:18.960 { 00:10:18.960 "job": "raid_bdev1", 00:10:18.960 "core_mask": "0x1", 00:10:18.960 "workload": "randrw", 00:10:18.960 "percentage": 50, 00:10:18.960 "status": "finished", 00:10:18.960 "queue_depth": 1, 00:10:18.960 "io_size": 131072, 00:10:18.960 "runtime": 1.384704, 00:10:18.960 "iops": 17007.967045664634, 00:10:18.960 "mibps": 2125.9958807080793, 00:10:18.960 "io_failed": 1, 00:10:18.960 "io_timeout": 0, 00:10:18.960 "avg_latency_us": 81.57948547560281, 00:10:18.960 "min_latency_us": 24.593886462882097, 00:10:18.960 "max_latency_us": 1309.2890829694322 00:10:18.960 } 00:10:18.960 ], 00:10:18.960 "core_count": 1 00:10:18.960 } 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83528 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 83528 ']' 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 83528 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83528 00:10:18.960 killing process with pid 83528 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83528' 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 83528 00:10:18.960 [2024-09-29 16:48:40.581497] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:18.960 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 83528 00:10:18.960 [2024-09-29 16:48:40.616360] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:19.220 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.wR9y7vROk1 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:19.221 ************************************ 00:10:19.221 END TEST raid_write_error_test 00:10:19.221 ************************************ 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:10:19.221 00:10:19.221 real 0m3.335s 00:10:19.221 user 0m4.201s 00:10:19.221 sys 0m0.514s 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:19.221 16:48:40 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.482 16:48:40 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:19.482 16:48:40 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:10:19.482 16:48:40 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:19.482 16:48:40 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:19.482 16:48:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:19.482 ************************************ 00:10:19.482 START TEST raid_state_function_test 00:10:19.482 ************************************ 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 false 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=83655 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83655' 00:10:19.482 Process raid pid: 83655 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 83655 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 83655 ']' 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:19.482 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:19.482 16:48:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.482 [2024-09-29 16:48:41.027909] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:19.482 [2024-09-29 16:48:41.028029] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:19.743 [2024-09-29 16:48:41.172214] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:19.743 [2024-09-29 16:48:41.217556] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:19.743 [2024-09-29 16:48:41.258951] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:19.743 [2024-09-29 16:48:41.258992] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.312 [2024-09-29 16:48:41.847908] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:20.312 [2024-09-29 16:48:41.847958] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:20.312 [2024-09-29 16:48:41.847969] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:20.312 [2024-09-29 16:48:41.847979] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:20.312 [2024-09-29 16:48:41.847985] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:20.312 [2024-09-29 16:48:41.847996] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:20.312 [2024-09-29 16:48:41.848002] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:20.312 [2024-09-29 16:48:41.848012] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.312 "name": "Existed_Raid", 00:10:20.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.312 "strip_size_kb": 0, 00:10:20.312 "state": "configuring", 00:10:20.312 "raid_level": "raid1", 00:10:20.312 "superblock": false, 00:10:20.312 "num_base_bdevs": 4, 00:10:20.312 "num_base_bdevs_discovered": 0, 00:10:20.312 "num_base_bdevs_operational": 4, 00:10:20.312 "base_bdevs_list": [ 00:10:20.312 { 00:10:20.312 "name": "BaseBdev1", 00:10:20.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.312 "is_configured": false, 00:10:20.312 "data_offset": 0, 00:10:20.312 "data_size": 0 00:10:20.312 }, 00:10:20.312 { 00:10:20.312 "name": "BaseBdev2", 00:10:20.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.312 "is_configured": false, 00:10:20.312 "data_offset": 0, 00:10:20.312 "data_size": 0 00:10:20.312 }, 00:10:20.312 { 00:10:20.312 "name": "BaseBdev3", 00:10:20.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.312 "is_configured": false, 00:10:20.312 "data_offset": 0, 00:10:20.312 "data_size": 0 00:10:20.312 }, 00:10:20.312 { 00:10:20.312 "name": "BaseBdev4", 00:10:20.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.312 "is_configured": false, 00:10:20.312 "data_offset": 0, 00:10:20.312 "data_size": 0 00:10:20.312 } 00:10:20.312 ] 00:10:20.312 }' 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.312 16:48:41 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.572 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:20.572 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.572 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.832 [2024-09-29 16:48:42.251109] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:20.832 [2024-09-29 16:48:42.251191] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.832 [2024-09-29 16:48:42.263110] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:20.832 [2024-09-29 16:48:42.263184] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:20.832 [2024-09-29 16:48:42.263210] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:20.832 [2024-09-29 16:48:42.263232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:20.832 [2024-09-29 16:48:42.263249] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:20.832 [2024-09-29 16:48:42.263269] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:20.832 [2024-09-29 16:48:42.263286] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:20.832 [2024-09-29 16:48:42.263306] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.832 [2024-09-29 16:48:42.283989] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:20.832 BaseBdev1 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.832 [ 00:10:20.832 { 00:10:20.832 "name": "BaseBdev1", 00:10:20.832 "aliases": [ 00:10:20.832 "56fbba27-fcde-4bd5-980d-a6430dc8cee7" 00:10:20.832 ], 00:10:20.832 "product_name": "Malloc disk", 00:10:20.832 "block_size": 512, 00:10:20.832 "num_blocks": 65536, 00:10:20.832 "uuid": "56fbba27-fcde-4bd5-980d-a6430dc8cee7", 00:10:20.832 "assigned_rate_limits": { 00:10:20.832 "rw_ios_per_sec": 0, 00:10:20.832 "rw_mbytes_per_sec": 0, 00:10:20.832 "r_mbytes_per_sec": 0, 00:10:20.832 "w_mbytes_per_sec": 0 00:10:20.832 }, 00:10:20.832 "claimed": true, 00:10:20.832 "claim_type": "exclusive_write", 00:10:20.832 "zoned": false, 00:10:20.832 "supported_io_types": { 00:10:20.832 "read": true, 00:10:20.832 "write": true, 00:10:20.832 "unmap": true, 00:10:20.832 "flush": true, 00:10:20.832 "reset": true, 00:10:20.832 "nvme_admin": false, 00:10:20.832 "nvme_io": false, 00:10:20.832 "nvme_io_md": false, 00:10:20.832 "write_zeroes": true, 00:10:20.832 "zcopy": true, 00:10:20.832 "get_zone_info": false, 00:10:20.832 "zone_management": false, 00:10:20.832 "zone_append": false, 00:10:20.832 "compare": false, 00:10:20.832 "compare_and_write": false, 00:10:20.832 "abort": true, 00:10:20.832 "seek_hole": false, 00:10:20.832 "seek_data": false, 00:10:20.832 "copy": true, 00:10:20.832 "nvme_iov_md": false 00:10:20.832 }, 00:10:20.832 "memory_domains": [ 00:10:20.832 { 00:10:20.832 "dma_device_id": "system", 00:10:20.832 "dma_device_type": 1 00:10:20.832 }, 00:10:20.832 { 00:10:20.832 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.832 "dma_device_type": 2 00:10:20.832 } 00:10:20.832 ], 00:10:20.832 "driver_specific": {} 00:10:20.832 } 00:10:20.832 ] 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:20.832 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:20.833 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:20.833 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.833 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.833 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.833 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:20.833 "name": "Existed_Raid", 00:10:20.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.833 "strip_size_kb": 0, 00:10:20.833 "state": "configuring", 00:10:20.833 "raid_level": "raid1", 00:10:20.833 "superblock": false, 00:10:20.833 "num_base_bdevs": 4, 00:10:20.833 "num_base_bdevs_discovered": 1, 00:10:20.833 "num_base_bdevs_operational": 4, 00:10:20.833 "base_bdevs_list": [ 00:10:20.833 { 00:10:20.833 "name": "BaseBdev1", 00:10:20.833 "uuid": "56fbba27-fcde-4bd5-980d-a6430dc8cee7", 00:10:20.833 "is_configured": true, 00:10:20.833 "data_offset": 0, 00:10:20.833 "data_size": 65536 00:10:20.833 }, 00:10:20.833 { 00:10:20.833 "name": "BaseBdev2", 00:10:20.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.833 "is_configured": false, 00:10:20.833 "data_offset": 0, 00:10:20.833 "data_size": 0 00:10:20.833 }, 00:10:20.833 { 00:10:20.833 "name": "BaseBdev3", 00:10:20.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.833 "is_configured": false, 00:10:20.833 "data_offset": 0, 00:10:20.833 "data_size": 0 00:10:20.833 }, 00:10:20.833 { 00:10:20.833 "name": "BaseBdev4", 00:10:20.833 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:20.833 "is_configured": false, 00:10:20.833 "data_offset": 0, 00:10:20.833 "data_size": 0 00:10:20.833 } 00:10:20.833 ] 00:10:20.833 }' 00:10:20.833 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:20.833 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.402 [2024-09-29 16:48:42.807134] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:21.402 [2024-09-29 16:48:42.807231] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.402 [2024-09-29 16:48:42.819151] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:21.402 [2024-09-29 16:48:42.820996] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:21.402 [2024-09-29 16:48:42.821079] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:21.402 [2024-09-29 16:48:42.821101] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:21.402 [2024-09-29 16:48:42.821111] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:21.402 [2024-09-29 16:48:42.821117] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:21.402 [2024-09-29 16:48:42.821125] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.402 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.403 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.403 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.403 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.403 "name": "Existed_Raid", 00:10:21.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.403 "strip_size_kb": 0, 00:10:21.403 "state": "configuring", 00:10:21.403 "raid_level": "raid1", 00:10:21.403 "superblock": false, 00:10:21.403 "num_base_bdevs": 4, 00:10:21.403 "num_base_bdevs_discovered": 1, 00:10:21.403 "num_base_bdevs_operational": 4, 00:10:21.403 "base_bdevs_list": [ 00:10:21.403 { 00:10:21.403 "name": "BaseBdev1", 00:10:21.403 "uuid": "56fbba27-fcde-4bd5-980d-a6430dc8cee7", 00:10:21.403 "is_configured": true, 00:10:21.403 "data_offset": 0, 00:10:21.403 "data_size": 65536 00:10:21.403 }, 00:10:21.403 { 00:10:21.403 "name": "BaseBdev2", 00:10:21.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.403 "is_configured": false, 00:10:21.403 "data_offset": 0, 00:10:21.403 "data_size": 0 00:10:21.403 }, 00:10:21.403 { 00:10:21.403 "name": "BaseBdev3", 00:10:21.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.403 "is_configured": false, 00:10:21.403 "data_offset": 0, 00:10:21.403 "data_size": 0 00:10:21.403 }, 00:10:21.403 { 00:10:21.403 "name": "BaseBdev4", 00:10:21.403 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.403 "is_configured": false, 00:10:21.403 "data_offset": 0, 00:10:21.403 "data_size": 0 00:10:21.403 } 00:10:21.403 ] 00:10:21.403 }' 00:10:21.403 16:48:42 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.403 16:48:42 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.662 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:21.662 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.663 [2024-09-29 16:48:43.257786] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:21.663 BaseBdev2 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.663 [ 00:10:21.663 { 00:10:21.663 "name": "BaseBdev2", 00:10:21.663 "aliases": [ 00:10:21.663 "8b834c00-4eff-4373-8fe0-446cf463bcff" 00:10:21.663 ], 00:10:21.663 "product_name": "Malloc disk", 00:10:21.663 "block_size": 512, 00:10:21.663 "num_blocks": 65536, 00:10:21.663 "uuid": "8b834c00-4eff-4373-8fe0-446cf463bcff", 00:10:21.663 "assigned_rate_limits": { 00:10:21.663 "rw_ios_per_sec": 0, 00:10:21.663 "rw_mbytes_per_sec": 0, 00:10:21.663 "r_mbytes_per_sec": 0, 00:10:21.663 "w_mbytes_per_sec": 0 00:10:21.663 }, 00:10:21.663 "claimed": true, 00:10:21.663 "claim_type": "exclusive_write", 00:10:21.663 "zoned": false, 00:10:21.663 "supported_io_types": { 00:10:21.663 "read": true, 00:10:21.663 "write": true, 00:10:21.663 "unmap": true, 00:10:21.663 "flush": true, 00:10:21.663 "reset": true, 00:10:21.663 "nvme_admin": false, 00:10:21.663 "nvme_io": false, 00:10:21.663 "nvme_io_md": false, 00:10:21.663 "write_zeroes": true, 00:10:21.663 "zcopy": true, 00:10:21.663 "get_zone_info": false, 00:10:21.663 "zone_management": false, 00:10:21.663 "zone_append": false, 00:10:21.663 "compare": false, 00:10:21.663 "compare_and_write": false, 00:10:21.663 "abort": true, 00:10:21.663 "seek_hole": false, 00:10:21.663 "seek_data": false, 00:10:21.663 "copy": true, 00:10:21.663 "nvme_iov_md": false 00:10:21.663 }, 00:10:21.663 "memory_domains": [ 00:10:21.663 { 00:10:21.663 "dma_device_id": "system", 00:10:21.663 "dma_device_type": 1 00:10:21.663 }, 00:10:21.663 { 00:10:21.663 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:21.663 "dma_device_type": 2 00:10:21.663 } 00:10:21.663 ], 00:10:21.663 "driver_specific": {} 00:10:21.663 } 00:10:21.663 ] 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:21.663 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:21.923 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:21.923 "name": "Existed_Raid", 00:10:21.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.923 "strip_size_kb": 0, 00:10:21.923 "state": "configuring", 00:10:21.923 "raid_level": "raid1", 00:10:21.923 "superblock": false, 00:10:21.923 "num_base_bdevs": 4, 00:10:21.923 "num_base_bdevs_discovered": 2, 00:10:21.923 "num_base_bdevs_operational": 4, 00:10:21.923 "base_bdevs_list": [ 00:10:21.923 { 00:10:21.923 "name": "BaseBdev1", 00:10:21.923 "uuid": "56fbba27-fcde-4bd5-980d-a6430dc8cee7", 00:10:21.923 "is_configured": true, 00:10:21.923 "data_offset": 0, 00:10:21.923 "data_size": 65536 00:10:21.923 }, 00:10:21.923 { 00:10:21.923 "name": "BaseBdev2", 00:10:21.923 "uuid": "8b834c00-4eff-4373-8fe0-446cf463bcff", 00:10:21.923 "is_configured": true, 00:10:21.923 "data_offset": 0, 00:10:21.923 "data_size": 65536 00:10:21.923 }, 00:10:21.923 { 00:10:21.923 "name": "BaseBdev3", 00:10:21.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.923 "is_configured": false, 00:10:21.923 "data_offset": 0, 00:10:21.923 "data_size": 0 00:10:21.923 }, 00:10:21.923 { 00:10:21.923 "name": "BaseBdev4", 00:10:21.923 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:21.923 "is_configured": false, 00:10:21.923 "data_offset": 0, 00:10:21.923 "data_size": 0 00:10:21.923 } 00:10:21.923 ] 00:10:21.923 }' 00:10:21.923 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:21.923 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.183 [2024-09-29 16:48:43.735748] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:22.183 BaseBdev3 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.183 [ 00:10:22.183 { 00:10:22.183 "name": "BaseBdev3", 00:10:22.183 "aliases": [ 00:10:22.183 "cc0bda19-eeba-4960-b89f-337fa72b94a2" 00:10:22.183 ], 00:10:22.183 "product_name": "Malloc disk", 00:10:22.183 "block_size": 512, 00:10:22.183 "num_blocks": 65536, 00:10:22.183 "uuid": "cc0bda19-eeba-4960-b89f-337fa72b94a2", 00:10:22.183 "assigned_rate_limits": { 00:10:22.183 "rw_ios_per_sec": 0, 00:10:22.183 "rw_mbytes_per_sec": 0, 00:10:22.183 "r_mbytes_per_sec": 0, 00:10:22.183 "w_mbytes_per_sec": 0 00:10:22.183 }, 00:10:22.183 "claimed": true, 00:10:22.183 "claim_type": "exclusive_write", 00:10:22.183 "zoned": false, 00:10:22.183 "supported_io_types": { 00:10:22.183 "read": true, 00:10:22.183 "write": true, 00:10:22.183 "unmap": true, 00:10:22.183 "flush": true, 00:10:22.183 "reset": true, 00:10:22.183 "nvme_admin": false, 00:10:22.183 "nvme_io": false, 00:10:22.183 "nvme_io_md": false, 00:10:22.183 "write_zeroes": true, 00:10:22.183 "zcopy": true, 00:10:22.183 "get_zone_info": false, 00:10:22.183 "zone_management": false, 00:10:22.183 "zone_append": false, 00:10:22.183 "compare": false, 00:10:22.183 "compare_and_write": false, 00:10:22.183 "abort": true, 00:10:22.183 "seek_hole": false, 00:10:22.183 "seek_data": false, 00:10:22.183 "copy": true, 00:10:22.183 "nvme_iov_md": false 00:10:22.183 }, 00:10:22.183 "memory_domains": [ 00:10:22.183 { 00:10:22.183 "dma_device_id": "system", 00:10:22.183 "dma_device_type": 1 00:10:22.183 }, 00:10:22.183 { 00:10:22.183 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.183 "dma_device_type": 2 00:10:22.183 } 00:10:22.183 ], 00:10:22.183 "driver_specific": {} 00:10:22.183 } 00:10:22.183 ] 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.183 "name": "Existed_Raid", 00:10:22.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.183 "strip_size_kb": 0, 00:10:22.183 "state": "configuring", 00:10:22.183 "raid_level": "raid1", 00:10:22.183 "superblock": false, 00:10:22.183 "num_base_bdevs": 4, 00:10:22.183 "num_base_bdevs_discovered": 3, 00:10:22.183 "num_base_bdevs_operational": 4, 00:10:22.183 "base_bdevs_list": [ 00:10:22.183 { 00:10:22.183 "name": "BaseBdev1", 00:10:22.183 "uuid": "56fbba27-fcde-4bd5-980d-a6430dc8cee7", 00:10:22.183 "is_configured": true, 00:10:22.183 "data_offset": 0, 00:10:22.183 "data_size": 65536 00:10:22.183 }, 00:10:22.183 { 00:10:22.183 "name": "BaseBdev2", 00:10:22.183 "uuid": "8b834c00-4eff-4373-8fe0-446cf463bcff", 00:10:22.183 "is_configured": true, 00:10:22.183 "data_offset": 0, 00:10:22.183 "data_size": 65536 00:10:22.183 }, 00:10:22.183 { 00:10:22.183 "name": "BaseBdev3", 00:10:22.183 "uuid": "cc0bda19-eeba-4960-b89f-337fa72b94a2", 00:10:22.183 "is_configured": true, 00:10:22.183 "data_offset": 0, 00:10:22.183 "data_size": 65536 00:10:22.183 }, 00:10:22.183 { 00:10:22.183 "name": "BaseBdev4", 00:10:22.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:22.183 "is_configured": false, 00:10:22.183 "data_offset": 0, 00:10:22.183 "data_size": 0 00:10:22.183 } 00:10:22.183 ] 00:10:22.183 }' 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.183 16:48:43 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.443 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:22.443 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.443 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.702 [2024-09-29 16:48:44.118047] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:22.702 [2024-09-29 16:48:44.118155] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:22.702 [2024-09-29 16:48:44.118181] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:22.702 [2024-09-29 16:48:44.118522] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:22.702 [2024-09-29 16:48:44.118731] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:22.702 [2024-09-29 16:48:44.118793] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:22.702 [2024-09-29 16:48:44.119068] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:22.702 BaseBdev4 00:10:22.702 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.702 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.703 [ 00:10:22.703 { 00:10:22.703 "name": "BaseBdev4", 00:10:22.703 "aliases": [ 00:10:22.703 "4a66712e-540a-4713-bc8c-1b158eeedaa8" 00:10:22.703 ], 00:10:22.703 "product_name": "Malloc disk", 00:10:22.703 "block_size": 512, 00:10:22.703 "num_blocks": 65536, 00:10:22.703 "uuid": "4a66712e-540a-4713-bc8c-1b158eeedaa8", 00:10:22.703 "assigned_rate_limits": { 00:10:22.703 "rw_ios_per_sec": 0, 00:10:22.703 "rw_mbytes_per_sec": 0, 00:10:22.703 "r_mbytes_per_sec": 0, 00:10:22.703 "w_mbytes_per_sec": 0 00:10:22.703 }, 00:10:22.703 "claimed": true, 00:10:22.703 "claim_type": "exclusive_write", 00:10:22.703 "zoned": false, 00:10:22.703 "supported_io_types": { 00:10:22.703 "read": true, 00:10:22.703 "write": true, 00:10:22.703 "unmap": true, 00:10:22.703 "flush": true, 00:10:22.703 "reset": true, 00:10:22.703 "nvme_admin": false, 00:10:22.703 "nvme_io": false, 00:10:22.703 "nvme_io_md": false, 00:10:22.703 "write_zeroes": true, 00:10:22.703 "zcopy": true, 00:10:22.703 "get_zone_info": false, 00:10:22.703 "zone_management": false, 00:10:22.703 "zone_append": false, 00:10:22.703 "compare": false, 00:10:22.703 "compare_and_write": false, 00:10:22.703 "abort": true, 00:10:22.703 "seek_hole": false, 00:10:22.703 "seek_data": false, 00:10:22.703 "copy": true, 00:10:22.703 "nvme_iov_md": false 00:10:22.703 }, 00:10:22.703 "memory_domains": [ 00:10:22.703 { 00:10:22.703 "dma_device_id": "system", 00:10:22.703 "dma_device_type": 1 00:10:22.703 }, 00:10:22.703 { 00:10:22.703 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:22.703 "dma_device_type": 2 00:10:22.703 } 00:10:22.703 ], 00:10:22.703 "driver_specific": {} 00:10:22.703 } 00:10:22.703 ] 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.703 "name": "Existed_Raid", 00:10:22.703 "uuid": "9dbcfacd-491e-4314-9f41-fc5aab72cfc3", 00:10:22.703 "strip_size_kb": 0, 00:10:22.703 "state": "online", 00:10:22.703 "raid_level": "raid1", 00:10:22.703 "superblock": false, 00:10:22.703 "num_base_bdevs": 4, 00:10:22.703 "num_base_bdevs_discovered": 4, 00:10:22.703 "num_base_bdevs_operational": 4, 00:10:22.703 "base_bdevs_list": [ 00:10:22.703 { 00:10:22.703 "name": "BaseBdev1", 00:10:22.703 "uuid": "56fbba27-fcde-4bd5-980d-a6430dc8cee7", 00:10:22.703 "is_configured": true, 00:10:22.703 "data_offset": 0, 00:10:22.703 "data_size": 65536 00:10:22.703 }, 00:10:22.703 { 00:10:22.703 "name": "BaseBdev2", 00:10:22.703 "uuid": "8b834c00-4eff-4373-8fe0-446cf463bcff", 00:10:22.703 "is_configured": true, 00:10:22.703 "data_offset": 0, 00:10:22.703 "data_size": 65536 00:10:22.703 }, 00:10:22.703 { 00:10:22.703 "name": "BaseBdev3", 00:10:22.703 "uuid": "cc0bda19-eeba-4960-b89f-337fa72b94a2", 00:10:22.703 "is_configured": true, 00:10:22.703 "data_offset": 0, 00:10:22.703 "data_size": 65536 00:10:22.703 }, 00:10:22.703 { 00:10:22.703 "name": "BaseBdev4", 00:10:22.703 "uuid": "4a66712e-540a-4713-bc8c-1b158eeedaa8", 00:10:22.703 "is_configured": true, 00:10:22.703 "data_offset": 0, 00:10:22.703 "data_size": 65536 00:10:22.703 } 00:10:22.703 ] 00:10:22.703 }' 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.703 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:22.963 [2024-09-29 16:48:44.613509] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:22.963 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.222 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:23.222 "name": "Existed_Raid", 00:10:23.222 "aliases": [ 00:10:23.222 "9dbcfacd-491e-4314-9f41-fc5aab72cfc3" 00:10:23.222 ], 00:10:23.222 "product_name": "Raid Volume", 00:10:23.222 "block_size": 512, 00:10:23.222 "num_blocks": 65536, 00:10:23.222 "uuid": "9dbcfacd-491e-4314-9f41-fc5aab72cfc3", 00:10:23.223 "assigned_rate_limits": { 00:10:23.223 "rw_ios_per_sec": 0, 00:10:23.223 "rw_mbytes_per_sec": 0, 00:10:23.223 "r_mbytes_per_sec": 0, 00:10:23.223 "w_mbytes_per_sec": 0 00:10:23.223 }, 00:10:23.223 "claimed": false, 00:10:23.223 "zoned": false, 00:10:23.223 "supported_io_types": { 00:10:23.223 "read": true, 00:10:23.223 "write": true, 00:10:23.223 "unmap": false, 00:10:23.223 "flush": false, 00:10:23.223 "reset": true, 00:10:23.223 "nvme_admin": false, 00:10:23.223 "nvme_io": false, 00:10:23.223 "nvme_io_md": false, 00:10:23.223 "write_zeroes": true, 00:10:23.223 "zcopy": false, 00:10:23.223 "get_zone_info": false, 00:10:23.223 "zone_management": false, 00:10:23.223 "zone_append": false, 00:10:23.223 "compare": false, 00:10:23.223 "compare_and_write": false, 00:10:23.223 "abort": false, 00:10:23.223 "seek_hole": false, 00:10:23.223 "seek_data": false, 00:10:23.223 "copy": false, 00:10:23.223 "nvme_iov_md": false 00:10:23.223 }, 00:10:23.223 "memory_domains": [ 00:10:23.223 { 00:10:23.223 "dma_device_id": "system", 00:10:23.223 "dma_device_type": 1 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:23.223 "dma_device_type": 2 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "dma_device_id": "system", 00:10:23.223 "dma_device_type": 1 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:23.223 "dma_device_type": 2 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "dma_device_id": "system", 00:10:23.223 "dma_device_type": 1 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:23.223 "dma_device_type": 2 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "dma_device_id": "system", 00:10:23.223 "dma_device_type": 1 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:23.223 "dma_device_type": 2 00:10:23.223 } 00:10:23.223 ], 00:10:23.223 "driver_specific": { 00:10:23.223 "raid": { 00:10:23.223 "uuid": "9dbcfacd-491e-4314-9f41-fc5aab72cfc3", 00:10:23.223 "strip_size_kb": 0, 00:10:23.223 "state": "online", 00:10:23.223 "raid_level": "raid1", 00:10:23.223 "superblock": false, 00:10:23.223 "num_base_bdevs": 4, 00:10:23.223 "num_base_bdevs_discovered": 4, 00:10:23.223 "num_base_bdevs_operational": 4, 00:10:23.223 "base_bdevs_list": [ 00:10:23.223 { 00:10:23.223 "name": "BaseBdev1", 00:10:23.223 "uuid": "56fbba27-fcde-4bd5-980d-a6430dc8cee7", 00:10:23.223 "is_configured": true, 00:10:23.223 "data_offset": 0, 00:10:23.223 "data_size": 65536 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "name": "BaseBdev2", 00:10:23.223 "uuid": "8b834c00-4eff-4373-8fe0-446cf463bcff", 00:10:23.223 "is_configured": true, 00:10:23.223 "data_offset": 0, 00:10:23.223 "data_size": 65536 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "name": "BaseBdev3", 00:10:23.223 "uuid": "cc0bda19-eeba-4960-b89f-337fa72b94a2", 00:10:23.223 "is_configured": true, 00:10:23.223 "data_offset": 0, 00:10:23.223 "data_size": 65536 00:10:23.223 }, 00:10:23.223 { 00:10:23.223 "name": "BaseBdev4", 00:10:23.223 "uuid": "4a66712e-540a-4713-bc8c-1b158eeedaa8", 00:10:23.223 "is_configured": true, 00:10:23.223 "data_offset": 0, 00:10:23.223 "data_size": 65536 00:10:23.223 } 00:10:23.223 ] 00:10:23.223 } 00:10:23.223 } 00:10:23.223 }' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:23.223 BaseBdev2 00:10:23.223 BaseBdev3 00:10:23.223 BaseBdev4' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:23.223 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.482 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.483 [2024-09-29 16:48:44.908762] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.483 "name": "Existed_Raid", 00:10:23.483 "uuid": "9dbcfacd-491e-4314-9f41-fc5aab72cfc3", 00:10:23.483 "strip_size_kb": 0, 00:10:23.483 "state": "online", 00:10:23.483 "raid_level": "raid1", 00:10:23.483 "superblock": false, 00:10:23.483 "num_base_bdevs": 4, 00:10:23.483 "num_base_bdevs_discovered": 3, 00:10:23.483 "num_base_bdevs_operational": 3, 00:10:23.483 "base_bdevs_list": [ 00:10:23.483 { 00:10:23.483 "name": null, 00:10:23.483 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:23.483 "is_configured": false, 00:10:23.483 "data_offset": 0, 00:10:23.483 "data_size": 65536 00:10:23.483 }, 00:10:23.483 { 00:10:23.483 "name": "BaseBdev2", 00:10:23.483 "uuid": "8b834c00-4eff-4373-8fe0-446cf463bcff", 00:10:23.483 "is_configured": true, 00:10:23.483 "data_offset": 0, 00:10:23.483 "data_size": 65536 00:10:23.483 }, 00:10:23.483 { 00:10:23.483 "name": "BaseBdev3", 00:10:23.483 "uuid": "cc0bda19-eeba-4960-b89f-337fa72b94a2", 00:10:23.483 "is_configured": true, 00:10:23.483 "data_offset": 0, 00:10:23.483 "data_size": 65536 00:10:23.483 }, 00:10:23.483 { 00:10:23.483 "name": "BaseBdev4", 00:10:23.483 "uuid": "4a66712e-540a-4713-bc8c-1b158eeedaa8", 00:10:23.483 "is_configured": true, 00:10:23.483 "data_offset": 0, 00:10:23.483 "data_size": 65536 00:10:23.483 } 00:10:23.483 ] 00:10:23.483 }' 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.483 16:48:44 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:23.742 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.743 [2024-09-29 16:48:45.383236] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.743 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 [2024-09-29 16:48:45.446155] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 [2024-09-29 16:48:45.517236] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:24.002 [2024-09-29 16:48:45.517321] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:24.002 [2024-09-29 16:48:45.528860] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:24.002 [2024-09-29 16:48:45.528909] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:24.002 [2024-09-29 16:48:45.528928] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 BaseBdev2 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.002 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.002 [ 00:10:24.002 { 00:10:24.002 "name": "BaseBdev2", 00:10:24.002 "aliases": [ 00:10:24.002 "e859cd5f-4914-4f87-af61-12a3d222ecc7" 00:10:24.002 ], 00:10:24.002 "product_name": "Malloc disk", 00:10:24.002 "block_size": 512, 00:10:24.002 "num_blocks": 65536, 00:10:24.002 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:24.002 "assigned_rate_limits": { 00:10:24.002 "rw_ios_per_sec": 0, 00:10:24.002 "rw_mbytes_per_sec": 0, 00:10:24.002 "r_mbytes_per_sec": 0, 00:10:24.002 "w_mbytes_per_sec": 0 00:10:24.002 }, 00:10:24.002 "claimed": false, 00:10:24.002 "zoned": false, 00:10:24.002 "supported_io_types": { 00:10:24.002 "read": true, 00:10:24.002 "write": true, 00:10:24.002 "unmap": true, 00:10:24.002 "flush": true, 00:10:24.002 "reset": true, 00:10:24.002 "nvme_admin": false, 00:10:24.002 "nvme_io": false, 00:10:24.002 "nvme_io_md": false, 00:10:24.002 "write_zeroes": true, 00:10:24.002 "zcopy": true, 00:10:24.003 "get_zone_info": false, 00:10:24.003 "zone_management": false, 00:10:24.003 "zone_append": false, 00:10:24.003 "compare": false, 00:10:24.003 "compare_and_write": false, 00:10:24.003 "abort": true, 00:10:24.003 "seek_hole": false, 00:10:24.003 "seek_data": false, 00:10:24.003 "copy": true, 00:10:24.003 "nvme_iov_md": false 00:10:24.003 }, 00:10:24.003 "memory_domains": [ 00:10:24.003 { 00:10:24.003 "dma_device_id": "system", 00:10:24.003 "dma_device_type": 1 00:10:24.003 }, 00:10:24.003 { 00:10:24.003 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:24.003 "dma_device_type": 2 00:10:24.003 } 00:10:24.003 ], 00:10:24.003 "driver_specific": {} 00:10:24.003 } 00:10:24.003 ] 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.003 BaseBdev3 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.003 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.262 [ 00:10:24.262 { 00:10:24.262 "name": "BaseBdev3", 00:10:24.262 "aliases": [ 00:10:24.262 "bc4b9f26-50e6-4786-b381-88c47e85e6f1" 00:10:24.262 ], 00:10:24.262 "product_name": "Malloc disk", 00:10:24.262 "block_size": 512, 00:10:24.262 "num_blocks": 65536, 00:10:24.262 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:24.262 "assigned_rate_limits": { 00:10:24.262 "rw_ios_per_sec": 0, 00:10:24.262 "rw_mbytes_per_sec": 0, 00:10:24.262 "r_mbytes_per_sec": 0, 00:10:24.262 "w_mbytes_per_sec": 0 00:10:24.262 }, 00:10:24.262 "claimed": false, 00:10:24.262 "zoned": false, 00:10:24.262 "supported_io_types": { 00:10:24.262 "read": true, 00:10:24.262 "write": true, 00:10:24.262 "unmap": true, 00:10:24.262 "flush": true, 00:10:24.262 "reset": true, 00:10:24.262 "nvme_admin": false, 00:10:24.262 "nvme_io": false, 00:10:24.262 "nvme_io_md": false, 00:10:24.262 "write_zeroes": true, 00:10:24.262 "zcopy": true, 00:10:24.262 "get_zone_info": false, 00:10:24.262 "zone_management": false, 00:10:24.262 "zone_append": false, 00:10:24.262 "compare": false, 00:10:24.262 "compare_and_write": false, 00:10:24.262 "abort": true, 00:10:24.262 "seek_hole": false, 00:10:24.262 "seek_data": false, 00:10:24.262 "copy": true, 00:10:24.262 "nvme_iov_md": false 00:10:24.262 }, 00:10:24.262 "memory_domains": [ 00:10:24.262 { 00:10:24.262 "dma_device_id": "system", 00:10:24.262 "dma_device_type": 1 00:10:24.262 }, 00:10:24.262 { 00:10:24.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:24.262 "dma_device_type": 2 00:10:24.262 } 00:10:24.262 ], 00:10:24.262 "driver_specific": {} 00:10:24.262 } 00:10:24.262 ] 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.262 BaseBdev4 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.262 [ 00:10:24.262 { 00:10:24.262 "name": "BaseBdev4", 00:10:24.262 "aliases": [ 00:10:24.262 "f3f62443-123f-4260-8e57-4054aa1a8022" 00:10:24.262 ], 00:10:24.262 "product_name": "Malloc disk", 00:10:24.262 "block_size": 512, 00:10:24.262 "num_blocks": 65536, 00:10:24.262 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:24.262 "assigned_rate_limits": { 00:10:24.262 "rw_ios_per_sec": 0, 00:10:24.262 "rw_mbytes_per_sec": 0, 00:10:24.262 "r_mbytes_per_sec": 0, 00:10:24.262 "w_mbytes_per_sec": 0 00:10:24.262 }, 00:10:24.262 "claimed": false, 00:10:24.262 "zoned": false, 00:10:24.262 "supported_io_types": { 00:10:24.262 "read": true, 00:10:24.262 "write": true, 00:10:24.262 "unmap": true, 00:10:24.262 "flush": true, 00:10:24.262 "reset": true, 00:10:24.262 "nvme_admin": false, 00:10:24.262 "nvme_io": false, 00:10:24.262 "nvme_io_md": false, 00:10:24.262 "write_zeroes": true, 00:10:24.262 "zcopy": true, 00:10:24.262 "get_zone_info": false, 00:10:24.262 "zone_management": false, 00:10:24.262 "zone_append": false, 00:10:24.262 "compare": false, 00:10:24.262 "compare_and_write": false, 00:10:24.262 "abort": true, 00:10:24.262 "seek_hole": false, 00:10:24.262 "seek_data": false, 00:10:24.262 "copy": true, 00:10:24.262 "nvme_iov_md": false 00:10:24.262 }, 00:10:24.262 "memory_domains": [ 00:10:24.262 { 00:10:24.262 "dma_device_id": "system", 00:10:24.262 "dma_device_type": 1 00:10:24.262 }, 00:10:24.262 { 00:10:24.262 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:24.262 "dma_device_type": 2 00:10:24.262 } 00:10:24.262 ], 00:10:24.262 "driver_specific": {} 00:10:24.262 } 00:10:24.262 ] 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.262 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.262 [2024-09-29 16:48:45.748859] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:24.262 [2024-09-29 16:48:45.748957] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:24.263 [2024-09-29 16:48:45.748995] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:24.263 [2024-09-29 16:48:45.750817] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:24.263 [2024-09-29 16:48:45.750903] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.263 "name": "Existed_Raid", 00:10:24.263 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.263 "strip_size_kb": 0, 00:10:24.263 "state": "configuring", 00:10:24.263 "raid_level": "raid1", 00:10:24.263 "superblock": false, 00:10:24.263 "num_base_bdevs": 4, 00:10:24.263 "num_base_bdevs_discovered": 3, 00:10:24.263 "num_base_bdevs_operational": 4, 00:10:24.263 "base_bdevs_list": [ 00:10:24.263 { 00:10:24.263 "name": "BaseBdev1", 00:10:24.263 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.263 "is_configured": false, 00:10:24.263 "data_offset": 0, 00:10:24.263 "data_size": 0 00:10:24.263 }, 00:10:24.263 { 00:10:24.263 "name": "BaseBdev2", 00:10:24.263 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:24.263 "is_configured": true, 00:10:24.263 "data_offset": 0, 00:10:24.263 "data_size": 65536 00:10:24.263 }, 00:10:24.263 { 00:10:24.263 "name": "BaseBdev3", 00:10:24.263 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:24.263 "is_configured": true, 00:10:24.263 "data_offset": 0, 00:10:24.263 "data_size": 65536 00:10:24.263 }, 00:10:24.263 { 00:10:24.263 "name": "BaseBdev4", 00:10:24.263 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:24.263 "is_configured": true, 00:10:24.263 "data_offset": 0, 00:10:24.263 "data_size": 65536 00:10:24.263 } 00:10:24.263 ] 00:10:24.263 }' 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.263 16:48:45 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.830 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.831 [2024-09-29 16:48:46.200069] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:24.831 "name": "Existed_Raid", 00:10:24.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.831 "strip_size_kb": 0, 00:10:24.831 "state": "configuring", 00:10:24.831 "raid_level": "raid1", 00:10:24.831 "superblock": false, 00:10:24.831 "num_base_bdevs": 4, 00:10:24.831 "num_base_bdevs_discovered": 2, 00:10:24.831 "num_base_bdevs_operational": 4, 00:10:24.831 "base_bdevs_list": [ 00:10:24.831 { 00:10:24.831 "name": "BaseBdev1", 00:10:24.831 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:24.831 "is_configured": false, 00:10:24.831 "data_offset": 0, 00:10:24.831 "data_size": 0 00:10:24.831 }, 00:10:24.831 { 00:10:24.831 "name": null, 00:10:24.831 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:24.831 "is_configured": false, 00:10:24.831 "data_offset": 0, 00:10:24.831 "data_size": 65536 00:10:24.831 }, 00:10:24.831 { 00:10:24.831 "name": "BaseBdev3", 00:10:24.831 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:24.831 "is_configured": true, 00:10:24.831 "data_offset": 0, 00:10:24.831 "data_size": 65536 00:10:24.831 }, 00:10:24.831 { 00:10:24.831 "name": "BaseBdev4", 00:10:24.831 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:24.831 "is_configured": true, 00:10:24.831 "data_offset": 0, 00:10:24.831 "data_size": 65536 00:10:24.831 } 00:10:24.831 ] 00:10:24.831 }' 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:24.831 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.090 [2024-09-29 16:48:46.670158] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:25.090 BaseBdev1 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.090 [ 00:10:25.090 { 00:10:25.090 "name": "BaseBdev1", 00:10:25.090 "aliases": [ 00:10:25.090 "3e930729-66b2-4fdd-a41c-a9cffee274dd" 00:10:25.090 ], 00:10:25.090 "product_name": "Malloc disk", 00:10:25.090 "block_size": 512, 00:10:25.090 "num_blocks": 65536, 00:10:25.090 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:25.090 "assigned_rate_limits": { 00:10:25.090 "rw_ios_per_sec": 0, 00:10:25.090 "rw_mbytes_per_sec": 0, 00:10:25.090 "r_mbytes_per_sec": 0, 00:10:25.090 "w_mbytes_per_sec": 0 00:10:25.090 }, 00:10:25.090 "claimed": true, 00:10:25.090 "claim_type": "exclusive_write", 00:10:25.090 "zoned": false, 00:10:25.090 "supported_io_types": { 00:10:25.090 "read": true, 00:10:25.090 "write": true, 00:10:25.090 "unmap": true, 00:10:25.090 "flush": true, 00:10:25.090 "reset": true, 00:10:25.090 "nvme_admin": false, 00:10:25.090 "nvme_io": false, 00:10:25.090 "nvme_io_md": false, 00:10:25.090 "write_zeroes": true, 00:10:25.090 "zcopy": true, 00:10:25.090 "get_zone_info": false, 00:10:25.090 "zone_management": false, 00:10:25.090 "zone_append": false, 00:10:25.090 "compare": false, 00:10:25.090 "compare_and_write": false, 00:10:25.090 "abort": true, 00:10:25.090 "seek_hole": false, 00:10:25.090 "seek_data": false, 00:10:25.090 "copy": true, 00:10:25.090 "nvme_iov_md": false 00:10:25.090 }, 00:10:25.090 "memory_domains": [ 00:10:25.090 { 00:10:25.090 "dma_device_id": "system", 00:10:25.090 "dma_device_type": 1 00:10:25.090 }, 00:10:25.090 { 00:10:25.090 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:25.090 "dma_device_type": 2 00:10:25.090 } 00:10:25.090 ], 00:10:25.090 "driver_specific": {} 00:10:25.090 } 00:10:25.090 ] 00:10:25.090 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.091 "name": "Existed_Raid", 00:10:25.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.091 "strip_size_kb": 0, 00:10:25.091 "state": "configuring", 00:10:25.091 "raid_level": "raid1", 00:10:25.091 "superblock": false, 00:10:25.091 "num_base_bdevs": 4, 00:10:25.091 "num_base_bdevs_discovered": 3, 00:10:25.091 "num_base_bdevs_operational": 4, 00:10:25.091 "base_bdevs_list": [ 00:10:25.091 { 00:10:25.091 "name": "BaseBdev1", 00:10:25.091 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:25.091 "is_configured": true, 00:10:25.091 "data_offset": 0, 00:10:25.091 "data_size": 65536 00:10:25.091 }, 00:10:25.091 { 00:10:25.091 "name": null, 00:10:25.091 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:25.091 "is_configured": false, 00:10:25.091 "data_offset": 0, 00:10:25.091 "data_size": 65536 00:10:25.091 }, 00:10:25.091 { 00:10:25.091 "name": "BaseBdev3", 00:10:25.091 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:25.091 "is_configured": true, 00:10:25.091 "data_offset": 0, 00:10:25.091 "data_size": 65536 00:10:25.091 }, 00:10:25.091 { 00:10:25.091 "name": "BaseBdev4", 00:10:25.091 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:25.091 "is_configured": true, 00:10:25.091 "data_offset": 0, 00:10:25.091 "data_size": 65536 00:10:25.091 } 00:10:25.091 ] 00:10:25.091 }' 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.091 16:48:46 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.659 [2024-09-29 16:48:47.193304] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.659 "name": "Existed_Raid", 00:10:25.659 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:25.659 "strip_size_kb": 0, 00:10:25.659 "state": "configuring", 00:10:25.659 "raid_level": "raid1", 00:10:25.659 "superblock": false, 00:10:25.659 "num_base_bdevs": 4, 00:10:25.659 "num_base_bdevs_discovered": 2, 00:10:25.659 "num_base_bdevs_operational": 4, 00:10:25.659 "base_bdevs_list": [ 00:10:25.659 { 00:10:25.659 "name": "BaseBdev1", 00:10:25.659 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:25.659 "is_configured": true, 00:10:25.659 "data_offset": 0, 00:10:25.659 "data_size": 65536 00:10:25.659 }, 00:10:25.659 { 00:10:25.659 "name": null, 00:10:25.659 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:25.659 "is_configured": false, 00:10:25.659 "data_offset": 0, 00:10:25.659 "data_size": 65536 00:10:25.659 }, 00:10:25.659 { 00:10:25.659 "name": null, 00:10:25.659 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:25.659 "is_configured": false, 00:10:25.659 "data_offset": 0, 00:10:25.659 "data_size": 65536 00:10:25.659 }, 00:10:25.659 { 00:10:25.659 "name": "BaseBdev4", 00:10:25.659 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:25.659 "is_configured": true, 00:10:25.659 "data_offset": 0, 00:10:25.659 "data_size": 65536 00:10:25.659 } 00:10:25.659 ] 00:10:25.659 }' 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.659 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.227 [2024-09-29 16:48:47.716431] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.227 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.227 "name": "Existed_Raid", 00:10:26.227 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.227 "strip_size_kb": 0, 00:10:26.227 "state": "configuring", 00:10:26.227 "raid_level": "raid1", 00:10:26.227 "superblock": false, 00:10:26.227 "num_base_bdevs": 4, 00:10:26.227 "num_base_bdevs_discovered": 3, 00:10:26.227 "num_base_bdevs_operational": 4, 00:10:26.228 "base_bdevs_list": [ 00:10:26.228 { 00:10:26.228 "name": "BaseBdev1", 00:10:26.228 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:26.228 "is_configured": true, 00:10:26.228 "data_offset": 0, 00:10:26.228 "data_size": 65536 00:10:26.228 }, 00:10:26.228 { 00:10:26.228 "name": null, 00:10:26.228 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:26.228 "is_configured": false, 00:10:26.228 "data_offset": 0, 00:10:26.228 "data_size": 65536 00:10:26.228 }, 00:10:26.228 { 00:10:26.228 "name": "BaseBdev3", 00:10:26.228 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:26.228 "is_configured": true, 00:10:26.228 "data_offset": 0, 00:10:26.228 "data_size": 65536 00:10:26.228 }, 00:10:26.228 { 00:10:26.228 "name": "BaseBdev4", 00:10:26.228 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:26.228 "is_configured": true, 00:10:26.228 "data_offset": 0, 00:10:26.228 "data_size": 65536 00:10:26.228 } 00:10:26.228 ] 00:10:26.228 }' 00:10:26.228 16:48:47 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.228 16:48:47 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.485 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.485 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:26.485 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.486 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.486 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.744 [2024-09-29 16:48:48.183641] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:26.744 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:26.744 "name": "Existed_Raid", 00:10:26.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:26.744 "strip_size_kb": 0, 00:10:26.744 "state": "configuring", 00:10:26.744 "raid_level": "raid1", 00:10:26.744 "superblock": false, 00:10:26.744 "num_base_bdevs": 4, 00:10:26.744 "num_base_bdevs_discovered": 2, 00:10:26.744 "num_base_bdevs_operational": 4, 00:10:26.744 "base_bdevs_list": [ 00:10:26.744 { 00:10:26.744 "name": null, 00:10:26.744 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:26.744 "is_configured": false, 00:10:26.744 "data_offset": 0, 00:10:26.744 "data_size": 65536 00:10:26.744 }, 00:10:26.744 { 00:10:26.744 "name": null, 00:10:26.744 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:26.744 "is_configured": false, 00:10:26.744 "data_offset": 0, 00:10:26.744 "data_size": 65536 00:10:26.744 }, 00:10:26.744 { 00:10:26.745 "name": "BaseBdev3", 00:10:26.745 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:26.745 "is_configured": true, 00:10:26.745 "data_offset": 0, 00:10:26.745 "data_size": 65536 00:10:26.745 }, 00:10:26.745 { 00:10:26.745 "name": "BaseBdev4", 00:10:26.745 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:26.745 "is_configured": true, 00:10:26.745 "data_offset": 0, 00:10:26.745 "data_size": 65536 00:10:26.745 } 00:10:26.745 ] 00:10:26.745 }' 00:10:26.745 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:26.745 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.003 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:27.004 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.004 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.004 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.004 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.263 [2024-09-29 16:48:48.705410] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.263 "name": "Existed_Raid", 00:10:27.263 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:27.263 "strip_size_kb": 0, 00:10:27.263 "state": "configuring", 00:10:27.263 "raid_level": "raid1", 00:10:27.263 "superblock": false, 00:10:27.263 "num_base_bdevs": 4, 00:10:27.263 "num_base_bdevs_discovered": 3, 00:10:27.263 "num_base_bdevs_operational": 4, 00:10:27.263 "base_bdevs_list": [ 00:10:27.263 { 00:10:27.263 "name": null, 00:10:27.263 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:27.263 "is_configured": false, 00:10:27.263 "data_offset": 0, 00:10:27.263 "data_size": 65536 00:10:27.263 }, 00:10:27.263 { 00:10:27.263 "name": "BaseBdev2", 00:10:27.263 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:27.263 "is_configured": true, 00:10:27.263 "data_offset": 0, 00:10:27.263 "data_size": 65536 00:10:27.263 }, 00:10:27.263 { 00:10:27.263 "name": "BaseBdev3", 00:10:27.263 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:27.263 "is_configured": true, 00:10:27.263 "data_offset": 0, 00:10:27.263 "data_size": 65536 00:10:27.263 }, 00:10:27.263 { 00:10:27.263 "name": "BaseBdev4", 00:10:27.263 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:27.263 "is_configured": true, 00:10:27.263 "data_offset": 0, 00:10:27.263 "data_size": 65536 00:10:27.263 } 00:10:27.263 ] 00:10:27.263 }' 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.263 16:48:48 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 3e930729-66b2-4fdd-a41c-a9cffee274dd 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.521 [2024-09-29 16:48:49.183446] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:27.521 [2024-09-29 16:48:49.183555] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:27.521 [2024-09-29 16:48:49.183589] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:27.521 [2024-09-29 16:48:49.183905] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:27.521 [2024-09-29 16:48:49.184085] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:27.521 [2024-09-29 16:48:49.184128] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:27.521 [2024-09-29 16:48:49.184368] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:27.521 NewBaseBdev 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.521 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.779 [ 00:10:27.779 { 00:10:27.779 "name": "NewBaseBdev", 00:10:27.779 "aliases": [ 00:10:27.779 "3e930729-66b2-4fdd-a41c-a9cffee274dd" 00:10:27.779 ], 00:10:27.779 "product_name": "Malloc disk", 00:10:27.779 "block_size": 512, 00:10:27.779 "num_blocks": 65536, 00:10:27.779 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:27.779 "assigned_rate_limits": { 00:10:27.779 "rw_ios_per_sec": 0, 00:10:27.779 "rw_mbytes_per_sec": 0, 00:10:27.779 "r_mbytes_per_sec": 0, 00:10:27.779 "w_mbytes_per_sec": 0 00:10:27.779 }, 00:10:27.779 "claimed": true, 00:10:27.779 "claim_type": "exclusive_write", 00:10:27.779 "zoned": false, 00:10:27.779 "supported_io_types": { 00:10:27.779 "read": true, 00:10:27.779 "write": true, 00:10:27.779 "unmap": true, 00:10:27.779 "flush": true, 00:10:27.779 "reset": true, 00:10:27.779 "nvme_admin": false, 00:10:27.779 "nvme_io": false, 00:10:27.779 "nvme_io_md": false, 00:10:27.779 "write_zeroes": true, 00:10:27.779 "zcopy": true, 00:10:27.779 "get_zone_info": false, 00:10:27.779 "zone_management": false, 00:10:27.779 "zone_append": false, 00:10:27.779 "compare": false, 00:10:27.779 "compare_and_write": false, 00:10:27.779 "abort": true, 00:10:27.779 "seek_hole": false, 00:10:27.779 "seek_data": false, 00:10:27.779 "copy": true, 00:10:27.779 "nvme_iov_md": false 00:10:27.779 }, 00:10:27.779 "memory_domains": [ 00:10:27.779 { 00:10:27.779 "dma_device_id": "system", 00:10:27.779 "dma_device_type": 1 00:10:27.779 }, 00:10:27.779 { 00:10:27.779 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:27.779 "dma_device_type": 2 00:10:27.779 } 00:10:27.779 ], 00:10:27.779 "driver_specific": {} 00:10:27.779 } 00:10:27.779 ] 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.779 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.779 "name": "Existed_Raid", 00:10:27.779 "uuid": "20360a8b-da83-481b-b107-30180401242e", 00:10:27.779 "strip_size_kb": 0, 00:10:27.779 "state": "online", 00:10:27.779 "raid_level": "raid1", 00:10:27.779 "superblock": false, 00:10:27.779 "num_base_bdevs": 4, 00:10:27.779 "num_base_bdevs_discovered": 4, 00:10:27.779 "num_base_bdevs_operational": 4, 00:10:27.779 "base_bdevs_list": [ 00:10:27.779 { 00:10:27.779 "name": "NewBaseBdev", 00:10:27.779 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:27.779 "is_configured": true, 00:10:27.779 "data_offset": 0, 00:10:27.779 "data_size": 65536 00:10:27.779 }, 00:10:27.779 { 00:10:27.779 "name": "BaseBdev2", 00:10:27.779 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:27.779 "is_configured": true, 00:10:27.779 "data_offset": 0, 00:10:27.779 "data_size": 65536 00:10:27.779 }, 00:10:27.779 { 00:10:27.779 "name": "BaseBdev3", 00:10:27.779 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:27.780 "is_configured": true, 00:10:27.780 "data_offset": 0, 00:10:27.780 "data_size": 65536 00:10:27.780 }, 00:10:27.780 { 00:10:27.780 "name": "BaseBdev4", 00:10:27.780 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:27.780 "is_configured": true, 00:10:27.780 "data_offset": 0, 00:10:27.780 "data_size": 65536 00:10:27.780 } 00:10:27.780 ] 00:10:27.780 }' 00:10:27.780 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.780 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.039 [2024-09-29 16:48:49.643062] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:28.039 "name": "Existed_Raid", 00:10:28.039 "aliases": [ 00:10:28.039 "20360a8b-da83-481b-b107-30180401242e" 00:10:28.039 ], 00:10:28.039 "product_name": "Raid Volume", 00:10:28.039 "block_size": 512, 00:10:28.039 "num_blocks": 65536, 00:10:28.039 "uuid": "20360a8b-da83-481b-b107-30180401242e", 00:10:28.039 "assigned_rate_limits": { 00:10:28.039 "rw_ios_per_sec": 0, 00:10:28.039 "rw_mbytes_per_sec": 0, 00:10:28.039 "r_mbytes_per_sec": 0, 00:10:28.039 "w_mbytes_per_sec": 0 00:10:28.039 }, 00:10:28.039 "claimed": false, 00:10:28.039 "zoned": false, 00:10:28.039 "supported_io_types": { 00:10:28.039 "read": true, 00:10:28.039 "write": true, 00:10:28.039 "unmap": false, 00:10:28.039 "flush": false, 00:10:28.039 "reset": true, 00:10:28.039 "nvme_admin": false, 00:10:28.039 "nvme_io": false, 00:10:28.039 "nvme_io_md": false, 00:10:28.039 "write_zeroes": true, 00:10:28.039 "zcopy": false, 00:10:28.039 "get_zone_info": false, 00:10:28.039 "zone_management": false, 00:10:28.039 "zone_append": false, 00:10:28.039 "compare": false, 00:10:28.039 "compare_and_write": false, 00:10:28.039 "abort": false, 00:10:28.039 "seek_hole": false, 00:10:28.039 "seek_data": false, 00:10:28.039 "copy": false, 00:10:28.039 "nvme_iov_md": false 00:10:28.039 }, 00:10:28.039 "memory_domains": [ 00:10:28.039 { 00:10:28.039 "dma_device_id": "system", 00:10:28.039 "dma_device_type": 1 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.039 "dma_device_type": 2 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "dma_device_id": "system", 00:10:28.039 "dma_device_type": 1 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.039 "dma_device_type": 2 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "dma_device_id": "system", 00:10:28.039 "dma_device_type": 1 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.039 "dma_device_type": 2 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "dma_device_id": "system", 00:10:28.039 "dma_device_type": 1 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:28.039 "dma_device_type": 2 00:10:28.039 } 00:10:28.039 ], 00:10:28.039 "driver_specific": { 00:10:28.039 "raid": { 00:10:28.039 "uuid": "20360a8b-da83-481b-b107-30180401242e", 00:10:28.039 "strip_size_kb": 0, 00:10:28.039 "state": "online", 00:10:28.039 "raid_level": "raid1", 00:10:28.039 "superblock": false, 00:10:28.039 "num_base_bdevs": 4, 00:10:28.039 "num_base_bdevs_discovered": 4, 00:10:28.039 "num_base_bdevs_operational": 4, 00:10:28.039 "base_bdevs_list": [ 00:10:28.039 { 00:10:28.039 "name": "NewBaseBdev", 00:10:28.039 "uuid": "3e930729-66b2-4fdd-a41c-a9cffee274dd", 00:10:28.039 "is_configured": true, 00:10:28.039 "data_offset": 0, 00:10:28.039 "data_size": 65536 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "name": "BaseBdev2", 00:10:28.039 "uuid": "e859cd5f-4914-4f87-af61-12a3d222ecc7", 00:10:28.039 "is_configured": true, 00:10:28.039 "data_offset": 0, 00:10:28.039 "data_size": 65536 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "name": "BaseBdev3", 00:10:28.039 "uuid": "bc4b9f26-50e6-4786-b381-88c47e85e6f1", 00:10:28.039 "is_configured": true, 00:10:28.039 "data_offset": 0, 00:10:28.039 "data_size": 65536 00:10:28.039 }, 00:10:28.039 { 00:10:28.039 "name": "BaseBdev4", 00:10:28.039 "uuid": "f3f62443-123f-4260-8e57-4054aa1a8022", 00:10:28.039 "is_configured": true, 00:10:28.039 "data_offset": 0, 00:10:28.039 "data_size": 65536 00:10:28.039 } 00:10:28.039 ] 00:10:28.039 } 00:10:28.039 } 00:10:28.039 }' 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:28.039 BaseBdev2 00:10:28.039 BaseBdev3 00:10:28.039 BaseBdev4' 00:10:28.039 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.297 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.298 [2024-09-29 16:48:49.942228] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:28.298 [2024-09-29 16:48:49.942295] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:28.298 [2024-09-29 16:48:49.942397] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:28.298 [2024-09-29 16:48:49.942705] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:28.298 [2024-09-29 16:48:49.942788] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 83655 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 83655 ']' 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 83655 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:28.298 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83655 00:10:28.557 killing process with pid 83655 00:10:28.557 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:28.557 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:28.557 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83655' 00:10:28.557 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 83655 00:10:28.557 [2024-09-29 16:48:49.985122] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:28.557 16:48:49 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 83655 00:10:28.557 [2024-09-29 16:48:50.025022] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:28.816 16:48:50 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:28.816 00:10:28.816 real 0m9.329s 00:10:28.816 user 0m15.981s 00:10:28.816 sys 0m1.884s 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:28.817 ************************************ 00:10:28.817 END TEST raid_state_function_test 00:10:28.817 ************************************ 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.817 16:48:50 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:10:28.817 16:48:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:28.817 16:48:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:28.817 16:48:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:28.817 ************************************ 00:10:28.817 START TEST raid_state_function_test_sb 00:10:28.817 ************************************ 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 true 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84303 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84303' 00:10:28.817 Process raid pid: 84303 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84303 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 84303 ']' 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:28.817 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:28.817 16:48:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:28.817 [2024-09-29 16:48:50.435211] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:28.817 [2024-09-29 16:48:50.435432] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:29.075 [2024-09-29 16:48:50.562979] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:29.075 [2024-09-29 16:48:50.607510] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:29.075 [2024-09-29 16:48:50.649950] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:29.075 [2024-09-29 16:48:50.650030] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.645 [2024-09-29 16:48:51.267383] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:29.645 [2024-09-29 16:48:51.267438] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:29.645 [2024-09-29 16:48:51.267451] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:29.645 [2024-09-29 16:48:51.267462] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:29.645 [2024-09-29 16:48:51.267468] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:29.645 [2024-09-29 16:48:51.267479] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:29.645 [2024-09-29 16:48:51.267484] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:29.645 [2024-09-29 16:48:51.267492] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.645 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.905 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.906 "name": "Existed_Raid", 00:10:29.906 "uuid": "5aa35117-637f-4070-9e85-d57ea3bbf02f", 00:10:29.906 "strip_size_kb": 0, 00:10:29.906 "state": "configuring", 00:10:29.906 "raid_level": "raid1", 00:10:29.906 "superblock": true, 00:10:29.906 "num_base_bdevs": 4, 00:10:29.906 "num_base_bdevs_discovered": 0, 00:10:29.906 "num_base_bdevs_operational": 4, 00:10:29.906 "base_bdevs_list": [ 00:10:29.906 { 00:10:29.906 "name": "BaseBdev1", 00:10:29.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.906 "is_configured": false, 00:10:29.906 "data_offset": 0, 00:10:29.906 "data_size": 0 00:10:29.906 }, 00:10:29.906 { 00:10:29.906 "name": "BaseBdev2", 00:10:29.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.906 "is_configured": false, 00:10:29.906 "data_offset": 0, 00:10:29.906 "data_size": 0 00:10:29.906 }, 00:10:29.906 { 00:10:29.906 "name": "BaseBdev3", 00:10:29.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.906 "is_configured": false, 00:10:29.906 "data_offset": 0, 00:10:29.906 "data_size": 0 00:10:29.906 }, 00:10:29.906 { 00:10:29.906 "name": "BaseBdev4", 00:10:29.906 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.906 "is_configured": false, 00:10:29.906 "data_offset": 0, 00:10:29.906 "data_size": 0 00:10:29.906 } 00:10:29.906 ] 00:10:29.906 }' 00:10:29.906 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.906 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.169 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:30.169 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.169 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.169 [2024-09-29 16:48:51.682582] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:30.169 [2024-09-29 16:48:51.682620] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:30.169 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.169 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:30.169 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.169 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.169 [2024-09-29 16:48:51.690581] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:30.169 [2024-09-29 16:48:51.690652] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:30.169 [2024-09-29 16:48:51.690679] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:30.169 [2024-09-29 16:48:51.690700] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:30.169 [2024-09-29 16:48:51.690727] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:30.169 [2024-09-29 16:48:51.690748] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:30.169 [2024-09-29 16:48:51.690766] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:30.170 [2024-09-29 16:48:51.690827] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.170 [2024-09-29 16:48:51.707424] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:30.170 BaseBdev1 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.170 [ 00:10:30.170 { 00:10:30.170 "name": "BaseBdev1", 00:10:30.170 "aliases": [ 00:10:30.170 "a974221a-20d3-4232-bde3-d28ad9b3d577" 00:10:30.170 ], 00:10:30.170 "product_name": "Malloc disk", 00:10:30.170 "block_size": 512, 00:10:30.170 "num_blocks": 65536, 00:10:30.170 "uuid": "a974221a-20d3-4232-bde3-d28ad9b3d577", 00:10:30.170 "assigned_rate_limits": { 00:10:30.170 "rw_ios_per_sec": 0, 00:10:30.170 "rw_mbytes_per_sec": 0, 00:10:30.170 "r_mbytes_per_sec": 0, 00:10:30.170 "w_mbytes_per_sec": 0 00:10:30.170 }, 00:10:30.170 "claimed": true, 00:10:30.170 "claim_type": "exclusive_write", 00:10:30.170 "zoned": false, 00:10:30.170 "supported_io_types": { 00:10:30.170 "read": true, 00:10:30.170 "write": true, 00:10:30.170 "unmap": true, 00:10:30.170 "flush": true, 00:10:30.170 "reset": true, 00:10:30.170 "nvme_admin": false, 00:10:30.170 "nvme_io": false, 00:10:30.170 "nvme_io_md": false, 00:10:30.170 "write_zeroes": true, 00:10:30.170 "zcopy": true, 00:10:30.170 "get_zone_info": false, 00:10:30.170 "zone_management": false, 00:10:30.170 "zone_append": false, 00:10:30.170 "compare": false, 00:10:30.170 "compare_and_write": false, 00:10:30.170 "abort": true, 00:10:30.170 "seek_hole": false, 00:10:30.170 "seek_data": false, 00:10:30.170 "copy": true, 00:10:30.170 "nvme_iov_md": false 00:10:30.170 }, 00:10:30.170 "memory_domains": [ 00:10:30.170 { 00:10:30.170 "dma_device_id": "system", 00:10:30.170 "dma_device_type": 1 00:10:30.170 }, 00:10:30.170 { 00:10:30.170 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.170 "dma_device_type": 2 00:10:30.170 } 00:10:30.170 ], 00:10:30.170 "driver_specific": {} 00:10:30.170 } 00:10:30.170 ] 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.170 "name": "Existed_Raid", 00:10:30.170 "uuid": "9655ebd3-a6d6-474a-863a-96f81d3ed2a4", 00:10:30.170 "strip_size_kb": 0, 00:10:30.170 "state": "configuring", 00:10:30.170 "raid_level": "raid1", 00:10:30.170 "superblock": true, 00:10:30.170 "num_base_bdevs": 4, 00:10:30.170 "num_base_bdevs_discovered": 1, 00:10:30.170 "num_base_bdevs_operational": 4, 00:10:30.170 "base_bdevs_list": [ 00:10:30.170 { 00:10:30.170 "name": "BaseBdev1", 00:10:30.170 "uuid": "a974221a-20d3-4232-bde3-d28ad9b3d577", 00:10:30.170 "is_configured": true, 00:10:30.170 "data_offset": 2048, 00:10:30.170 "data_size": 63488 00:10:30.170 }, 00:10:30.170 { 00:10:30.170 "name": "BaseBdev2", 00:10:30.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.170 "is_configured": false, 00:10:30.170 "data_offset": 0, 00:10:30.170 "data_size": 0 00:10:30.170 }, 00:10:30.170 { 00:10:30.170 "name": "BaseBdev3", 00:10:30.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.170 "is_configured": false, 00:10:30.170 "data_offset": 0, 00:10:30.170 "data_size": 0 00:10:30.170 }, 00:10:30.170 { 00:10:30.170 "name": "BaseBdev4", 00:10:30.170 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.170 "is_configured": false, 00:10:30.170 "data_offset": 0, 00:10:30.170 "data_size": 0 00:10:30.170 } 00:10:30.170 ] 00:10:30.170 }' 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.170 16:48:51 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.752 [2024-09-29 16:48:52.206632] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:30.752 [2024-09-29 16:48:52.206740] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.752 [2024-09-29 16:48:52.218660] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:30.752 [2024-09-29 16:48:52.220525] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:30.752 [2024-09-29 16:48:52.220565] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:30.752 [2024-09-29 16:48:52.220574] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:30.752 [2024-09-29 16:48:52.220582] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:30.752 [2024-09-29 16:48:52.220588] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:30.752 [2024-09-29 16:48:52.220596] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.752 "name": "Existed_Raid", 00:10:30.752 "uuid": "0859d104-bd51-41d0-935e-89220496f20d", 00:10:30.752 "strip_size_kb": 0, 00:10:30.752 "state": "configuring", 00:10:30.752 "raid_level": "raid1", 00:10:30.752 "superblock": true, 00:10:30.752 "num_base_bdevs": 4, 00:10:30.752 "num_base_bdevs_discovered": 1, 00:10:30.752 "num_base_bdevs_operational": 4, 00:10:30.752 "base_bdevs_list": [ 00:10:30.752 { 00:10:30.752 "name": "BaseBdev1", 00:10:30.752 "uuid": "a974221a-20d3-4232-bde3-d28ad9b3d577", 00:10:30.752 "is_configured": true, 00:10:30.752 "data_offset": 2048, 00:10:30.752 "data_size": 63488 00:10:30.752 }, 00:10:30.752 { 00:10:30.752 "name": "BaseBdev2", 00:10:30.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.752 "is_configured": false, 00:10:30.752 "data_offset": 0, 00:10:30.752 "data_size": 0 00:10:30.752 }, 00:10:30.752 { 00:10:30.752 "name": "BaseBdev3", 00:10:30.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.752 "is_configured": false, 00:10:30.752 "data_offset": 0, 00:10:30.752 "data_size": 0 00:10:30.752 }, 00:10:30.752 { 00:10:30.752 "name": "BaseBdev4", 00:10:30.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.752 "is_configured": false, 00:10:30.752 "data_offset": 0, 00:10:30.752 "data_size": 0 00:10:30.752 } 00:10:30.752 ] 00:10:30.752 }' 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.752 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.013 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:31.013 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.013 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.273 [2024-09-29 16:48:52.703325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:31.273 BaseBdev2 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.273 [ 00:10:31.273 { 00:10:31.273 "name": "BaseBdev2", 00:10:31.273 "aliases": [ 00:10:31.273 "02e2d574-fa2c-4c2b-ab1f-37b15fa3737b" 00:10:31.273 ], 00:10:31.273 "product_name": "Malloc disk", 00:10:31.273 "block_size": 512, 00:10:31.273 "num_blocks": 65536, 00:10:31.273 "uuid": "02e2d574-fa2c-4c2b-ab1f-37b15fa3737b", 00:10:31.273 "assigned_rate_limits": { 00:10:31.273 "rw_ios_per_sec": 0, 00:10:31.273 "rw_mbytes_per_sec": 0, 00:10:31.273 "r_mbytes_per_sec": 0, 00:10:31.273 "w_mbytes_per_sec": 0 00:10:31.273 }, 00:10:31.273 "claimed": true, 00:10:31.273 "claim_type": "exclusive_write", 00:10:31.273 "zoned": false, 00:10:31.273 "supported_io_types": { 00:10:31.273 "read": true, 00:10:31.273 "write": true, 00:10:31.273 "unmap": true, 00:10:31.273 "flush": true, 00:10:31.273 "reset": true, 00:10:31.273 "nvme_admin": false, 00:10:31.273 "nvme_io": false, 00:10:31.273 "nvme_io_md": false, 00:10:31.273 "write_zeroes": true, 00:10:31.273 "zcopy": true, 00:10:31.273 "get_zone_info": false, 00:10:31.273 "zone_management": false, 00:10:31.273 "zone_append": false, 00:10:31.273 "compare": false, 00:10:31.273 "compare_and_write": false, 00:10:31.273 "abort": true, 00:10:31.273 "seek_hole": false, 00:10:31.273 "seek_data": false, 00:10:31.273 "copy": true, 00:10:31.273 "nvme_iov_md": false 00:10:31.273 }, 00:10:31.273 "memory_domains": [ 00:10:31.273 { 00:10:31.273 "dma_device_id": "system", 00:10:31.273 "dma_device_type": 1 00:10:31.273 }, 00:10:31.273 { 00:10:31.273 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:31.273 "dma_device_type": 2 00:10:31.273 } 00:10:31.273 ], 00:10:31.273 "driver_specific": {} 00:10:31.273 } 00:10:31.273 ] 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:31.273 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.274 "name": "Existed_Raid", 00:10:31.274 "uuid": "0859d104-bd51-41d0-935e-89220496f20d", 00:10:31.274 "strip_size_kb": 0, 00:10:31.274 "state": "configuring", 00:10:31.274 "raid_level": "raid1", 00:10:31.274 "superblock": true, 00:10:31.274 "num_base_bdevs": 4, 00:10:31.274 "num_base_bdevs_discovered": 2, 00:10:31.274 "num_base_bdevs_operational": 4, 00:10:31.274 "base_bdevs_list": [ 00:10:31.274 { 00:10:31.274 "name": "BaseBdev1", 00:10:31.274 "uuid": "a974221a-20d3-4232-bde3-d28ad9b3d577", 00:10:31.274 "is_configured": true, 00:10:31.274 "data_offset": 2048, 00:10:31.274 "data_size": 63488 00:10:31.274 }, 00:10:31.274 { 00:10:31.274 "name": "BaseBdev2", 00:10:31.274 "uuid": "02e2d574-fa2c-4c2b-ab1f-37b15fa3737b", 00:10:31.274 "is_configured": true, 00:10:31.274 "data_offset": 2048, 00:10:31.274 "data_size": 63488 00:10:31.274 }, 00:10:31.274 { 00:10:31.274 "name": "BaseBdev3", 00:10:31.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.274 "is_configured": false, 00:10:31.274 "data_offset": 0, 00:10:31.274 "data_size": 0 00:10:31.274 }, 00:10:31.274 { 00:10:31.274 "name": "BaseBdev4", 00:10:31.274 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.274 "is_configured": false, 00:10:31.274 "data_offset": 0, 00:10:31.274 "data_size": 0 00:10:31.274 } 00:10:31.274 ] 00:10:31.274 }' 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.274 16:48:52 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.534 BaseBdev3 00:10:31.534 [2024-09-29 16:48:53.181509] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.534 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.797 [ 00:10:31.797 { 00:10:31.797 "name": "BaseBdev3", 00:10:31.797 "aliases": [ 00:10:31.797 "76a9ffb9-2dcc-4b0e-90e3-c97367fca511" 00:10:31.797 ], 00:10:31.797 "product_name": "Malloc disk", 00:10:31.797 "block_size": 512, 00:10:31.797 "num_blocks": 65536, 00:10:31.797 "uuid": "76a9ffb9-2dcc-4b0e-90e3-c97367fca511", 00:10:31.797 "assigned_rate_limits": { 00:10:31.797 "rw_ios_per_sec": 0, 00:10:31.797 "rw_mbytes_per_sec": 0, 00:10:31.797 "r_mbytes_per_sec": 0, 00:10:31.797 "w_mbytes_per_sec": 0 00:10:31.797 }, 00:10:31.797 "claimed": true, 00:10:31.797 "claim_type": "exclusive_write", 00:10:31.797 "zoned": false, 00:10:31.797 "supported_io_types": { 00:10:31.797 "read": true, 00:10:31.797 "write": true, 00:10:31.797 "unmap": true, 00:10:31.797 "flush": true, 00:10:31.797 "reset": true, 00:10:31.797 "nvme_admin": false, 00:10:31.797 "nvme_io": false, 00:10:31.797 "nvme_io_md": false, 00:10:31.797 "write_zeroes": true, 00:10:31.797 "zcopy": true, 00:10:31.797 "get_zone_info": false, 00:10:31.797 "zone_management": false, 00:10:31.797 "zone_append": false, 00:10:31.797 "compare": false, 00:10:31.797 "compare_and_write": false, 00:10:31.797 "abort": true, 00:10:31.797 "seek_hole": false, 00:10:31.797 "seek_data": false, 00:10:31.797 "copy": true, 00:10:31.797 "nvme_iov_md": false 00:10:31.797 }, 00:10:31.797 "memory_domains": [ 00:10:31.797 { 00:10:31.797 "dma_device_id": "system", 00:10:31.797 "dma_device_type": 1 00:10:31.797 }, 00:10:31.797 { 00:10:31.797 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:31.797 "dma_device_type": 2 00:10:31.797 } 00:10:31.797 ], 00:10:31.797 "driver_specific": {} 00:10:31.797 } 00:10:31.797 ] 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.797 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.797 "name": "Existed_Raid", 00:10:31.797 "uuid": "0859d104-bd51-41d0-935e-89220496f20d", 00:10:31.797 "strip_size_kb": 0, 00:10:31.797 "state": "configuring", 00:10:31.798 "raid_level": "raid1", 00:10:31.798 "superblock": true, 00:10:31.798 "num_base_bdevs": 4, 00:10:31.798 "num_base_bdevs_discovered": 3, 00:10:31.798 "num_base_bdevs_operational": 4, 00:10:31.798 "base_bdevs_list": [ 00:10:31.798 { 00:10:31.798 "name": "BaseBdev1", 00:10:31.798 "uuid": "a974221a-20d3-4232-bde3-d28ad9b3d577", 00:10:31.798 "is_configured": true, 00:10:31.798 "data_offset": 2048, 00:10:31.798 "data_size": 63488 00:10:31.798 }, 00:10:31.798 { 00:10:31.798 "name": "BaseBdev2", 00:10:31.798 "uuid": "02e2d574-fa2c-4c2b-ab1f-37b15fa3737b", 00:10:31.798 "is_configured": true, 00:10:31.798 "data_offset": 2048, 00:10:31.798 "data_size": 63488 00:10:31.798 }, 00:10:31.798 { 00:10:31.798 "name": "BaseBdev3", 00:10:31.798 "uuid": "76a9ffb9-2dcc-4b0e-90e3-c97367fca511", 00:10:31.798 "is_configured": true, 00:10:31.798 "data_offset": 2048, 00:10:31.798 "data_size": 63488 00:10:31.798 }, 00:10:31.798 { 00:10:31.798 "name": "BaseBdev4", 00:10:31.798 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.798 "is_configured": false, 00:10:31.798 "data_offset": 0, 00:10:31.798 "data_size": 0 00:10:31.798 } 00:10:31.798 ] 00:10:31.798 }' 00:10:31.798 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.798 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.058 BaseBdev4 00:10:32.058 [2024-09-29 16:48:53.608543] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:32.058 [2024-09-29 16:48:53.608758] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:32.058 [2024-09-29 16:48:53.608786] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:32.058 [2024-09-29 16:48:53.609060] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:32.058 [2024-09-29 16:48:53.609256] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:32.058 [2024-09-29 16:48:53.609326] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.058 [2024-09-29 16:48:53.609487] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.058 [ 00:10:32.058 { 00:10:32.058 "name": "BaseBdev4", 00:10:32.058 "aliases": [ 00:10:32.058 "93ddbb01-87a9-4543-bcdb-47165cd0204c" 00:10:32.058 ], 00:10:32.058 "product_name": "Malloc disk", 00:10:32.058 "block_size": 512, 00:10:32.058 "num_blocks": 65536, 00:10:32.058 "uuid": "93ddbb01-87a9-4543-bcdb-47165cd0204c", 00:10:32.058 "assigned_rate_limits": { 00:10:32.058 "rw_ios_per_sec": 0, 00:10:32.058 "rw_mbytes_per_sec": 0, 00:10:32.058 "r_mbytes_per_sec": 0, 00:10:32.058 "w_mbytes_per_sec": 0 00:10:32.058 }, 00:10:32.058 "claimed": true, 00:10:32.058 "claim_type": "exclusive_write", 00:10:32.058 "zoned": false, 00:10:32.058 "supported_io_types": { 00:10:32.058 "read": true, 00:10:32.058 "write": true, 00:10:32.058 "unmap": true, 00:10:32.058 "flush": true, 00:10:32.058 "reset": true, 00:10:32.058 "nvme_admin": false, 00:10:32.058 "nvme_io": false, 00:10:32.058 "nvme_io_md": false, 00:10:32.058 "write_zeroes": true, 00:10:32.058 "zcopy": true, 00:10:32.058 "get_zone_info": false, 00:10:32.058 "zone_management": false, 00:10:32.058 "zone_append": false, 00:10:32.058 "compare": false, 00:10:32.058 "compare_and_write": false, 00:10:32.058 "abort": true, 00:10:32.058 "seek_hole": false, 00:10:32.058 "seek_data": false, 00:10:32.058 "copy": true, 00:10:32.058 "nvme_iov_md": false 00:10:32.058 }, 00:10:32.058 "memory_domains": [ 00:10:32.058 { 00:10:32.058 "dma_device_id": "system", 00:10:32.058 "dma_device_type": 1 00:10:32.058 }, 00:10:32.058 { 00:10:32.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.058 "dma_device_type": 2 00:10:32.058 } 00:10:32.058 ], 00:10:32.058 "driver_specific": {} 00:10:32.058 } 00:10:32.058 ] 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.058 "name": "Existed_Raid", 00:10:32.058 "uuid": "0859d104-bd51-41d0-935e-89220496f20d", 00:10:32.058 "strip_size_kb": 0, 00:10:32.058 "state": "online", 00:10:32.058 "raid_level": "raid1", 00:10:32.058 "superblock": true, 00:10:32.058 "num_base_bdevs": 4, 00:10:32.058 "num_base_bdevs_discovered": 4, 00:10:32.058 "num_base_bdevs_operational": 4, 00:10:32.058 "base_bdevs_list": [ 00:10:32.058 { 00:10:32.058 "name": "BaseBdev1", 00:10:32.058 "uuid": "a974221a-20d3-4232-bde3-d28ad9b3d577", 00:10:32.058 "is_configured": true, 00:10:32.058 "data_offset": 2048, 00:10:32.058 "data_size": 63488 00:10:32.058 }, 00:10:32.058 { 00:10:32.058 "name": "BaseBdev2", 00:10:32.058 "uuid": "02e2d574-fa2c-4c2b-ab1f-37b15fa3737b", 00:10:32.058 "is_configured": true, 00:10:32.058 "data_offset": 2048, 00:10:32.058 "data_size": 63488 00:10:32.058 }, 00:10:32.058 { 00:10:32.058 "name": "BaseBdev3", 00:10:32.058 "uuid": "76a9ffb9-2dcc-4b0e-90e3-c97367fca511", 00:10:32.058 "is_configured": true, 00:10:32.058 "data_offset": 2048, 00:10:32.058 "data_size": 63488 00:10:32.058 }, 00:10:32.058 { 00:10:32.058 "name": "BaseBdev4", 00:10:32.058 "uuid": "93ddbb01-87a9-4543-bcdb-47165cd0204c", 00:10:32.058 "is_configured": true, 00:10:32.058 "data_offset": 2048, 00:10:32.058 "data_size": 63488 00:10:32.058 } 00:10:32.058 ] 00:10:32.058 }' 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.058 16:48:53 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.629 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:32.630 [2024-09-29 16:48:54.068136] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:32.630 "name": "Existed_Raid", 00:10:32.630 "aliases": [ 00:10:32.630 "0859d104-bd51-41d0-935e-89220496f20d" 00:10:32.630 ], 00:10:32.630 "product_name": "Raid Volume", 00:10:32.630 "block_size": 512, 00:10:32.630 "num_blocks": 63488, 00:10:32.630 "uuid": "0859d104-bd51-41d0-935e-89220496f20d", 00:10:32.630 "assigned_rate_limits": { 00:10:32.630 "rw_ios_per_sec": 0, 00:10:32.630 "rw_mbytes_per_sec": 0, 00:10:32.630 "r_mbytes_per_sec": 0, 00:10:32.630 "w_mbytes_per_sec": 0 00:10:32.630 }, 00:10:32.630 "claimed": false, 00:10:32.630 "zoned": false, 00:10:32.630 "supported_io_types": { 00:10:32.630 "read": true, 00:10:32.630 "write": true, 00:10:32.630 "unmap": false, 00:10:32.630 "flush": false, 00:10:32.630 "reset": true, 00:10:32.630 "nvme_admin": false, 00:10:32.630 "nvme_io": false, 00:10:32.630 "nvme_io_md": false, 00:10:32.630 "write_zeroes": true, 00:10:32.630 "zcopy": false, 00:10:32.630 "get_zone_info": false, 00:10:32.630 "zone_management": false, 00:10:32.630 "zone_append": false, 00:10:32.630 "compare": false, 00:10:32.630 "compare_and_write": false, 00:10:32.630 "abort": false, 00:10:32.630 "seek_hole": false, 00:10:32.630 "seek_data": false, 00:10:32.630 "copy": false, 00:10:32.630 "nvme_iov_md": false 00:10:32.630 }, 00:10:32.630 "memory_domains": [ 00:10:32.630 { 00:10:32.630 "dma_device_id": "system", 00:10:32.630 "dma_device_type": 1 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.630 "dma_device_type": 2 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "dma_device_id": "system", 00:10:32.630 "dma_device_type": 1 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.630 "dma_device_type": 2 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "dma_device_id": "system", 00:10:32.630 "dma_device_type": 1 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.630 "dma_device_type": 2 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "dma_device_id": "system", 00:10:32.630 "dma_device_type": 1 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.630 "dma_device_type": 2 00:10:32.630 } 00:10:32.630 ], 00:10:32.630 "driver_specific": { 00:10:32.630 "raid": { 00:10:32.630 "uuid": "0859d104-bd51-41d0-935e-89220496f20d", 00:10:32.630 "strip_size_kb": 0, 00:10:32.630 "state": "online", 00:10:32.630 "raid_level": "raid1", 00:10:32.630 "superblock": true, 00:10:32.630 "num_base_bdevs": 4, 00:10:32.630 "num_base_bdevs_discovered": 4, 00:10:32.630 "num_base_bdevs_operational": 4, 00:10:32.630 "base_bdevs_list": [ 00:10:32.630 { 00:10:32.630 "name": "BaseBdev1", 00:10:32.630 "uuid": "a974221a-20d3-4232-bde3-d28ad9b3d577", 00:10:32.630 "is_configured": true, 00:10:32.630 "data_offset": 2048, 00:10:32.630 "data_size": 63488 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "name": "BaseBdev2", 00:10:32.630 "uuid": "02e2d574-fa2c-4c2b-ab1f-37b15fa3737b", 00:10:32.630 "is_configured": true, 00:10:32.630 "data_offset": 2048, 00:10:32.630 "data_size": 63488 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "name": "BaseBdev3", 00:10:32.630 "uuid": "76a9ffb9-2dcc-4b0e-90e3-c97367fca511", 00:10:32.630 "is_configured": true, 00:10:32.630 "data_offset": 2048, 00:10:32.630 "data_size": 63488 00:10:32.630 }, 00:10:32.630 { 00:10:32.630 "name": "BaseBdev4", 00:10:32.630 "uuid": "93ddbb01-87a9-4543-bcdb-47165cd0204c", 00:10:32.630 "is_configured": true, 00:10:32.630 "data_offset": 2048, 00:10:32.630 "data_size": 63488 00:10:32.630 } 00:10:32.630 ] 00:10:32.630 } 00:10:32.630 } 00:10:32.630 }' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:32.630 BaseBdev2 00:10:32.630 BaseBdev3 00:10:32.630 BaseBdev4' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.630 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.890 [2024-09-29 16:48:54.375310] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.890 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.891 "name": "Existed_Raid", 00:10:32.891 "uuid": "0859d104-bd51-41d0-935e-89220496f20d", 00:10:32.891 "strip_size_kb": 0, 00:10:32.891 "state": "online", 00:10:32.891 "raid_level": "raid1", 00:10:32.891 "superblock": true, 00:10:32.891 "num_base_bdevs": 4, 00:10:32.891 "num_base_bdevs_discovered": 3, 00:10:32.891 "num_base_bdevs_operational": 3, 00:10:32.891 "base_bdevs_list": [ 00:10:32.891 { 00:10:32.891 "name": null, 00:10:32.891 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.891 "is_configured": false, 00:10:32.891 "data_offset": 0, 00:10:32.891 "data_size": 63488 00:10:32.891 }, 00:10:32.891 { 00:10:32.891 "name": "BaseBdev2", 00:10:32.891 "uuid": "02e2d574-fa2c-4c2b-ab1f-37b15fa3737b", 00:10:32.891 "is_configured": true, 00:10:32.891 "data_offset": 2048, 00:10:32.891 "data_size": 63488 00:10:32.891 }, 00:10:32.891 { 00:10:32.891 "name": "BaseBdev3", 00:10:32.891 "uuid": "76a9ffb9-2dcc-4b0e-90e3-c97367fca511", 00:10:32.891 "is_configured": true, 00:10:32.891 "data_offset": 2048, 00:10:32.891 "data_size": 63488 00:10:32.891 }, 00:10:32.891 { 00:10:32.891 "name": "BaseBdev4", 00:10:32.891 "uuid": "93ddbb01-87a9-4543-bcdb-47165cd0204c", 00:10:32.891 "is_configured": true, 00:10:32.891 "data_offset": 2048, 00:10:32.891 "data_size": 63488 00:10:32.891 } 00:10:32.891 ] 00:10:32.891 }' 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.891 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.151 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:33.151 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:33.151 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.151 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.151 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.151 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:33.151 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.411 [2024-09-29 16:48:54.837762] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.411 [2024-09-29 16:48:54.908885] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.411 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.412 [2024-09-29 16:48:54.963585] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:33.412 [2024-09-29 16:48:54.963740] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:33.412 [2024-09-29 16:48:54.975277] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:33.412 [2024-09-29 16:48:54.975327] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:33.412 [2024-09-29 16:48:54.975346] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.412 16:48:54 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.412 BaseBdev2 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.412 [ 00:10:33.412 { 00:10:33.412 "name": "BaseBdev2", 00:10:33.412 "aliases": [ 00:10:33.412 "2db53c81-c920-4faf-aaea-eb5f86547e57" 00:10:33.412 ], 00:10:33.412 "product_name": "Malloc disk", 00:10:33.412 "block_size": 512, 00:10:33.412 "num_blocks": 65536, 00:10:33.412 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:33.412 "assigned_rate_limits": { 00:10:33.412 "rw_ios_per_sec": 0, 00:10:33.412 "rw_mbytes_per_sec": 0, 00:10:33.412 "r_mbytes_per_sec": 0, 00:10:33.412 "w_mbytes_per_sec": 0 00:10:33.412 }, 00:10:33.412 "claimed": false, 00:10:33.412 "zoned": false, 00:10:33.412 "supported_io_types": { 00:10:33.412 "read": true, 00:10:33.412 "write": true, 00:10:33.412 "unmap": true, 00:10:33.412 "flush": true, 00:10:33.412 "reset": true, 00:10:33.412 "nvme_admin": false, 00:10:33.412 "nvme_io": false, 00:10:33.412 "nvme_io_md": false, 00:10:33.412 "write_zeroes": true, 00:10:33.412 "zcopy": true, 00:10:33.412 "get_zone_info": false, 00:10:33.412 "zone_management": false, 00:10:33.412 "zone_append": false, 00:10:33.412 "compare": false, 00:10:33.412 "compare_and_write": false, 00:10:33.412 "abort": true, 00:10:33.412 "seek_hole": false, 00:10:33.412 "seek_data": false, 00:10:33.412 "copy": true, 00:10:33.412 "nvme_iov_md": false 00:10:33.412 }, 00:10:33.412 "memory_domains": [ 00:10:33.412 { 00:10:33.412 "dma_device_id": "system", 00:10:33.412 "dma_device_type": 1 00:10:33.412 }, 00:10:33.412 { 00:10:33.412 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.412 "dma_device_type": 2 00:10:33.412 } 00:10:33.412 ], 00:10:33.412 "driver_specific": {} 00:10:33.412 } 00:10:33.412 ] 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.412 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.672 BaseBdev3 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.672 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.673 [ 00:10:33.673 { 00:10:33.673 "name": "BaseBdev3", 00:10:33.673 "aliases": [ 00:10:33.673 "5059030f-c0c0-4bd9-b786-7851b22753d6" 00:10:33.673 ], 00:10:33.673 "product_name": "Malloc disk", 00:10:33.673 "block_size": 512, 00:10:33.673 "num_blocks": 65536, 00:10:33.673 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:33.673 "assigned_rate_limits": { 00:10:33.673 "rw_ios_per_sec": 0, 00:10:33.673 "rw_mbytes_per_sec": 0, 00:10:33.673 "r_mbytes_per_sec": 0, 00:10:33.673 "w_mbytes_per_sec": 0 00:10:33.673 }, 00:10:33.673 "claimed": false, 00:10:33.673 "zoned": false, 00:10:33.673 "supported_io_types": { 00:10:33.673 "read": true, 00:10:33.673 "write": true, 00:10:33.673 "unmap": true, 00:10:33.673 "flush": true, 00:10:33.673 "reset": true, 00:10:33.673 "nvme_admin": false, 00:10:33.673 "nvme_io": false, 00:10:33.673 "nvme_io_md": false, 00:10:33.673 "write_zeroes": true, 00:10:33.673 "zcopy": true, 00:10:33.673 "get_zone_info": false, 00:10:33.673 "zone_management": false, 00:10:33.673 "zone_append": false, 00:10:33.673 "compare": false, 00:10:33.673 "compare_and_write": false, 00:10:33.673 "abort": true, 00:10:33.673 "seek_hole": false, 00:10:33.673 "seek_data": false, 00:10:33.673 "copy": true, 00:10:33.673 "nvme_iov_md": false 00:10:33.673 }, 00:10:33.673 "memory_domains": [ 00:10:33.673 { 00:10:33.673 "dma_device_id": "system", 00:10:33.673 "dma_device_type": 1 00:10:33.673 }, 00:10:33.673 { 00:10:33.673 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.673 "dma_device_type": 2 00:10:33.673 } 00:10:33.673 ], 00:10:33.673 "driver_specific": {} 00:10:33.673 } 00:10:33.673 ] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.673 BaseBdev4 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.673 [ 00:10:33.673 { 00:10:33.673 "name": "BaseBdev4", 00:10:33.673 "aliases": [ 00:10:33.673 "3c169108-45d7-4e85-b0a9-f819daab7d0f" 00:10:33.673 ], 00:10:33.673 "product_name": "Malloc disk", 00:10:33.673 "block_size": 512, 00:10:33.673 "num_blocks": 65536, 00:10:33.673 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:33.673 "assigned_rate_limits": { 00:10:33.673 "rw_ios_per_sec": 0, 00:10:33.673 "rw_mbytes_per_sec": 0, 00:10:33.673 "r_mbytes_per_sec": 0, 00:10:33.673 "w_mbytes_per_sec": 0 00:10:33.673 }, 00:10:33.673 "claimed": false, 00:10:33.673 "zoned": false, 00:10:33.673 "supported_io_types": { 00:10:33.673 "read": true, 00:10:33.673 "write": true, 00:10:33.673 "unmap": true, 00:10:33.673 "flush": true, 00:10:33.673 "reset": true, 00:10:33.673 "nvme_admin": false, 00:10:33.673 "nvme_io": false, 00:10:33.673 "nvme_io_md": false, 00:10:33.673 "write_zeroes": true, 00:10:33.673 "zcopy": true, 00:10:33.673 "get_zone_info": false, 00:10:33.673 "zone_management": false, 00:10:33.673 "zone_append": false, 00:10:33.673 "compare": false, 00:10:33.673 "compare_and_write": false, 00:10:33.673 "abort": true, 00:10:33.673 "seek_hole": false, 00:10:33.673 "seek_data": false, 00:10:33.673 "copy": true, 00:10:33.673 "nvme_iov_md": false 00:10:33.673 }, 00:10:33.673 "memory_domains": [ 00:10:33.673 { 00:10:33.673 "dma_device_id": "system", 00:10:33.673 "dma_device_type": 1 00:10:33.673 }, 00:10:33.673 { 00:10:33.673 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.673 "dma_device_type": 2 00:10:33.673 } 00:10:33.673 ], 00:10:33.673 "driver_specific": {} 00:10:33.673 } 00:10:33.673 ] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.673 [2024-09-29 16:48:55.178944] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:33.673 [2024-09-29 16:48:55.179043] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:33.673 [2024-09-29 16:48:55.179080] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:33.673 [2024-09-29 16:48:55.180896] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:33.673 [2024-09-29 16:48:55.180990] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.673 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.673 "name": "Existed_Raid", 00:10:33.673 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:33.673 "strip_size_kb": 0, 00:10:33.673 "state": "configuring", 00:10:33.673 "raid_level": "raid1", 00:10:33.673 "superblock": true, 00:10:33.673 "num_base_bdevs": 4, 00:10:33.673 "num_base_bdevs_discovered": 3, 00:10:33.673 "num_base_bdevs_operational": 4, 00:10:33.673 "base_bdevs_list": [ 00:10:33.673 { 00:10:33.673 "name": "BaseBdev1", 00:10:33.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.673 "is_configured": false, 00:10:33.673 "data_offset": 0, 00:10:33.673 "data_size": 0 00:10:33.673 }, 00:10:33.673 { 00:10:33.673 "name": "BaseBdev2", 00:10:33.673 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:33.673 "is_configured": true, 00:10:33.673 "data_offset": 2048, 00:10:33.673 "data_size": 63488 00:10:33.673 }, 00:10:33.673 { 00:10:33.673 "name": "BaseBdev3", 00:10:33.673 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:33.673 "is_configured": true, 00:10:33.673 "data_offset": 2048, 00:10:33.673 "data_size": 63488 00:10:33.673 }, 00:10:33.673 { 00:10:33.673 "name": "BaseBdev4", 00:10:33.673 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:33.673 "is_configured": true, 00:10:33.673 "data_offset": 2048, 00:10:33.673 "data_size": 63488 00:10:33.674 } 00:10:33.674 ] 00:10:33.674 }' 00:10:33.674 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.674 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.243 [2024-09-29 16:48:55.654098] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.243 "name": "Existed_Raid", 00:10:34.243 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:34.243 "strip_size_kb": 0, 00:10:34.243 "state": "configuring", 00:10:34.243 "raid_level": "raid1", 00:10:34.243 "superblock": true, 00:10:34.243 "num_base_bdevs": 4, 00:10:34.243 "num_base_bdevs_discovered": 2, 00:10:34.243 "num_base_bdevs_operational": 4, 00:10:34.243 "base_bdevs_list": [ 00:10:34.243 { 00:10:34.243 "name": "BaseBdev1", 00:10:34.243 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.243 "is_configured": false, 00:10:34.243 "data_offset": 0, 00:10:34.243 "data_size": 0 00:10:34.243 }, 00:10:34.243 { 00:10:34.243 "name": null, 00:10:34.243 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:34.243 "is_configured": false, 00:10:34.243 "data_offset": 0, 00:10:34.243 "data_size": 63488 00:10:34.243 }, 00:10:34.243 { 00:10:34.243 "name": "BaseBdev3", 00:10:34.243 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:34.243 "is_configured": true, 00:10:34.243 "data_offset": 2048, 00:10:34.243 "data_size": 63488 00:10:34.243 }, 00:10:34.243 { 00:10:34.243 "name": "BaseBdev4", 00:10:34.243 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:34.243 "is_configured": true, 00:10:34.243 "data_offset": 2048, 00:10:34.243 "data_size": 63488 00:10:34.243 } 00:10:34.243 ] 00:10:34.243 }' 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.243 16:48:55 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.503 BaseBdev1 00:10:34.503 [2024-09-29 16:48:56.112295] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.503 [ 00:10:34.503 { 00:10:34.503 "name": "BaseBdev1", 00:10:34.503 "aliases": [ 00:10:34.503 "0ea572f9-86f7-40de-a29a-f2eaf000875a" 00:10:34.503 ], 00:10:34.503 "product_name": "Malloc disk", 00:10:34.503 "block_size": 512, 00:10:34.503 "num_blocks": 65536, 00:10:34.503 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:34.503 "assigned_rate_limits": { 00:10:34.503 "rw_ios_per_sec": 0, 00:10:34.503 "rw_mbytes_per_sec": 0, 00:10:34.503 "r_mbytes_per_sec": 0, 00:10:34.503 "w_mbytes_per_sec": 0 00:10:34.503 }, 00:10:34.503 "claimed": true, 00:10:34.503 "claim_type": "exclusive_write", 00:10:34.503 "zoned": false, 00:10:34.503 "supported_io_types": { 00:10:34.503 "read": true, 00:10:34.503 "write": true, 00:10:34.503 "unmap": true, 00:10:34.503 "flush": true, 00:10:34.503 "reset": true, 00:10:34.503 "nvme_admin": false, 00:10:34.503 "nvme_io": false, 00:10:34.503 "nvme_io_md": false, 00:10:34.503 "write_zeroes": true, 00:10:34.503 "zcopy": true, 00:10:34.503 "get_zone_info": false, 00:10:34.503 "zone_management": false, 00:10:34.503 "zone_append": false, 00:10:34.503 "compare": false, 00:10:34.503 "compare_and_write": false, 00:10:34.503 "abort": true, 00:10:34.503 "seek_hole": false, 00:10:34.503 "seek_data": false, 00:10:34.503 "copy": true, 00:10:34.503 "nvme_iov_md": false 00:10:34.503 }, 00:10:34.503 "memory_domains": [ 00:10:34.503 { 00:10:34.503 "dma_device_id": "system", 00:10:34.503 "dma_device_type": 1 00:10:34.503 }, 00:10:34.503 { 00:10:34.503 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.503 "dma_device_type": 2 00:10:34.503 } 00:10:34.503 ], 00:10:34.503 "driver_specific": {} 00:10:34.503 } 00:10:34.503 ] 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.503 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:34.776 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.776 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.776 "name": "Existed_Raid", 00:10:34.776 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:34.776 "strip_size_kb": 0, 00:10:34.776 "state": "configuring", 00:10:34.776 "raid_level": "raid1", 00:10:34.776 "superblock": true, 00:10:34.776 "num_base_bdevs": 4, 00:10:34.776 "num_base_bdevs_discovered": 3, 00:10:34.776 "num_base_bdevs_operational": 4, 00:10:34.776 "base_bdevs_list": [ 00:10:34.776 { 00:10:34.776 "name": "BaseBdev1", 00:10:34.776 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:34.776 "is_configured": true, 00:10:34.776 "data_offset": 2048, 00:10:34.776 "data_size": 63488 00:10:34.776 }, 00:10:34.776 { 00:10:34.776 "name": null, 00:10:34.776 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:34.776 "is_configured": false, 00:10:34.776 "data_offset": 0, 00:10:34.776 "data_size": 63488 00:10:34.776 }, 00:10:34.776 { 00:10:34.776 "name": "BaseBdev3", 00:10:34.776 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:34.776 "is_configured": true, 00:10:34.776 "data_offset": 2048, 00:10:34.776 "data_size": 63488 00:10:34.776 }, 00:10:34.776 { 00:10:34.776 "name": "BaseBdev4", 00:10:34.776 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:34.776 "is_configured": true, 00:10:34.776 "data_offset": 2048, 00:10:34.776 "data_size": 63488 00:10:34.776 } 00:10:34.776 ] 00:10:34.776 }' 00:10:34.776 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.776 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.036 [2024-09-29 16:48:56.655390] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.036 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.037 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.296 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.296 "name": "Existed_Raid", 00:10:35.296 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:35.296 "strip_size_kb": 0, 00:10:35.296 "state": "configuring", 00:10:35.296 "raid_level": "raid1", 00:10:35.296 "superblock": true, 00:10:35.296 "num_base_bdevs": 4, 00:10:35.296 "num_base_bdevs_discovered": 2, 00:10:35.296 "num_base_bdevs_operational": 4, 00:10:35.296 "base_bdevs_list": [ 00:10:35.296 { 00:10:35.296 "name": "BaseBdev1", 00:10:35.296 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:35.296 "is_configured": true, 00:10:35.296 "data_offset": 2048, 00:10:35.296 "data_size": 63488 00:10:35.296 }, 00:10:35.296 { 00:10:35.296 "name": null, 00:10:35.296 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:35.296 "is_configured": false, 00:10:35.296 "data_offset": 0, 00:10:35.296 "data_size": 63488 00:10:35.296 }, 00:10:35.296 { 00:10:35.296 "name": null, 00:10:35.296 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:35.296 "is_configured": false, 00:10:35.296 "data_offset": 0, 00:10:35.296 "data_size": 63488 00:10:35.296 }, 00:10:35.296 { 00:10:35.296 "name": "BaseBdev4", 00:10:35.296 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:35.296 "is_configured": true, 00:10:35.296 "data_offset": 2048, 00:10:35.296 "data_size": 63488 00:10:35.296 } 00:10:35.296 ] 00:10:35.296 }' 00:10:35.296 16:48:56 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.296 16:48:56 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.556 [2024-09-29 16:48:57.178563] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:35.556 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.815 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.815 "name": "Existed_Raid", 00:10:35.815 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:35.815 "strip_size_kb": 0, 00:10:35.815 "state": "configuring", 00:10:35.815 "raid_level": "raid1", 00:10:35.815 "superblock": true, 00:10:35.815 "num_base_bdevs": 4, 00:10:35.815 "num_base_bdevs_discovered": 3, 00:10:35.815 "num_base_bdevs_operational": 4, 00:10:35.815 "base_bdevs_list": [ 00:10:35.815 { 00:10:35.815 "name": "BaseBdev1", 00:10:35.815 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:35.815 "is_configured": true, 00:10:35.815 "data_offset": 2048, 00:10:35.815 "data_size": 63488 00:10:35.815 }, 00:10:35.815 { 00:10:35.815 "name": null, 00:10:35.815 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:35.815 "is_configured": false, 00:10:35.815 "data_offset": 0, 00:10:35.815 "data_size": 63488 00:10:35.815 }, 00:10:35.815 { 00:10:35.815 "name": "BaseBdev3", 00:10:35.815 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:35.815 "is_configured": true, 00:10:35.815 "data_offset": 2048, 00:10:35.815 "data_size": 63488 00:10:35.815 }, 00:10:35.815 { 00:10:35.815 "name": "BaseBdev4", 00:10:35.815 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:35.815 "is_configured": true, 00:10:35.815 "data_offset": 2048, 00:10:35.815 "data_size": 63488 00:10:35.815 } 00:10:35.815 ] 00:10:35.815 }' 00:10:35.815 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.815 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.075 [2024-09-29 16:48:57.657727] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.075 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.076 "name": "Existed_Raid", 00:10:36.076 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:36.076 "strip_size_kb": 0, 00:10:36.076 "state": "configuring", 00:10:36.076 "raid_level": "raid1", 00:10:36.076 "superblock": true, 00:10:36.076 "num_base_bdevs": 4, 00:10:36.076 "num_base_bdevs_discovered": 2, 00:10:36.076 "num_base_bdevs_operational": 4, 00:10:36.076 "base_bdevs_list": [ 00:10:36.076 { 00:10:36.076 "name": null, 00:10:36.076 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:36.076 "is_configured": false, 00:10:36.076 "data_offset": 0, 00:10:36.076 "data_size": 63488 00:10:36.076 }, 00:10:36.076 { 00:10:36.076 "name": null, 00:10:36.076 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:36.076 "is_configured": false, 00:10:36.076 "data_offset": 0, 00:10:36.076 "data_size": 63488 00:10:36.076 }, 00:10:36.076 { 00:10:36.076 "name": "BaseBdev3", 00:10:36.076 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:36.076 "is_configured": true, 00:10:36.076 "data_offset": 2048, 00:10:36.076 "data_size": 63488 00:10:36.076 }, 00:10:36.076 { 00:10:36.076 "name": "BaseBdev4", 00:10:36.076 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:36.076 "is_configured": true, 00:10:36.076 "data_offset": 2048, 00:10:36.076 "data_size": 63488 00:10:36.076 } 00:10:36.076 ] 00:10:36.076 }' 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.076 16:48:57 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.645 [2024-09-29 16:48:58.159351] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.645 "name": "Existed_Raid", 00:10:36.645 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:36.645 "strip_size_kb": 0, 00:10:36.645 "state": "configuring", 00:10:36.645 "raid_level": "raid1", 00:10:36.645 "superblock": true, 00:10:36.645 "num_base_bdevs": 4, 00:10:36.645 "num_base_bdevs_discovered": 3, 00:10:36.645 "num_base_bdevs_operational": 4, 00:10:36.645 "base_bdevs_list": [ 00:10:36.645 { 00:10:36.645 "name": null, 00:10:36.645 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:36.645 "is_configured": false, 00:10:36.645 "data_offset": 0, 00:10:36.645 "data_size": 63488 00:10:36.645 }, 00:10:36.645 { 00:10:36.645 "name": "BaseBdev2", 00:10:36.645 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:36.645 "is_configured": true, 00:10:36.645 "data_offset": 2048, 00:10:36.645 "data_size": 63488 00:10:36.645 }, 00:10:36.645 { 00:10:36.645 "name": "BaseBdev3", 00:10:36.645 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:36.645 "is_configured": true, 00:10:36.645 "data_offset": 2048, 00:10:36.645 "data_size": 63488 00:10:36.645 }, 00:10:36.645 { 00:10:36.645 "name": "BaseBdev4", 00:10:36.645 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:36.645 "is_configured": true, 00:10:36.645 "data_offset": 2048, 00:10:36.645 "data_size": 63488 00:10:36.645 } 00:10:36.645 ] 00:10:36.645 }' 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.645 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 0ea572f9-86f7-40de-a29a-f2eaf000875a 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.215 [2024-09-29 16:48:58.677351] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:37.215 [2024-09-29 16:48:58.677623] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:37.215 NewBaseBdev 00:10:37.215 [2024-09-29 16:48:58.677675] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:37.215 [2024-09-29 16:48:58.677955] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:37.215 [2024-09-29 16:48:58.678074] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:37.215 [2024-09-29 16:48:58.678085] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:37.215 [2024-09-29 16:48:58.678184] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.215 [ 00:10:37.215 { 00:10:37.215 "name": "NewBaseBdev", 00:10:37.215 "aliases": [ 00:10:37.215 "0ea572f9-86f7-40de-a29a-f2eaf000875a" 00:10:37.215 ], 00:10:37.215 "product_name": "Malloc disk", 00:10:37.215 "block_size": 512, 00:10:37.215 "num_blocks": 65536, 00:10:37.215 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:37.215 "assigned_rate_limits": { 00:10:37.215 "rw_ios_per_sec": 0, 00:10:37.215 "rw_mbytes_per_sec": 0, 00:10:37.215 "r_mbytes_per_sec": 0, 00:10:37.215 "w_mbytes_per_sec": 0 00:10:37.215 }, 00:10:37.215 "claimed": true, 00:10:37.215 "claim_type": "exclusive_write", 00:10:37.215 "zoned": false, 00:10:37.215 "supported_io_types": { 00:10:37.215 "read": true, 00:10:37.215 "write": true, 00:10:37.215 "unmap": true, 00:10:37.215 "flush": true, 00:10:37.215 "reset": true, 00:10:37.215 "nvme_admin": false, 00:10:37.215 "nvme_io": false, 00:10:37.215 "nvme_io_md": false, 00:10:37.215 "write_zeroes": true, 00:10:37.215 "zcopy": true, 00:10:37.215 "get_zone_info": false, 00:10:37.215 "zone_management": false, 00:10:37.215 "zone_append": false, 00:10:37.215 "compare": false, 00:10:37.215 "compare_and_write": false, 00:10:37.215 "abort": true, 00:10:37.215 "seek_hole": false, 00:10:37.215 "seek_data": false, 00:10:37.215 "copy": true, 00:10:37.215 "nvme_iov_md": false 00:10:37.215 }, 00:10:37.215 "memory_domains": [ 00:10:37.215 { 00:10:37.215 "dma_device_id": "system", 00:10:37.215 "dma_device_type": 1 00:10:37.215 }, 00:10:37.215 { 00:10:37.215 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.215 "dma_device_type": 2 00:10:37.215 } 00:10:37.215 ], 00:10:37.215 "driver_specific": {} 00:10:37.215 } 00:10:37.215 ] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.215 "name": "Existed_Raid", 00:10:37.215 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:37.215 "strip_size_kb": 0, 00:10:37.215 "state": "online", 00:10:37.215 "raid_level": "raid1", 00:10:37.215 "superblock": true, 00:10:37.215 "num_base_bdevs": 4, 00:10:37.215 "num_base_bdevs_discovered": 4, 00:10:37.215 "num_base_bdevs_operational": 4, 00:10:37.215 "base_bdevs_list": [ 00:10:37.215 { 00:10:37.215 "name": "NewBaseBdev", 00:10:37.215 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:37.215 "is_configured": true, 00:10:37.215 "data_offset": 2048, 00:10:37.215 "data_size": 63488 00:10:37.215 }, 00:10:37.215 { 00:10:37.215 "name": "BaseBdev2", 00:10:37.215 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:37.215 "is_configured": true, 00:10:37.215 "data_offset": 2048, 00:10:37.215 "data_size": 63488 00:10:37.215 }, 00:10:37.215 { 00:10:37.215 "name": "BaseBdev3", 00:10:37.215 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:37.215 "is_configured": true, 00:10:37.215 "data_offset": 2048, 00:10:37.215 "data_size": 63488 00:10:37.215 }, 00:10:37.215 { 00:10:37.215 "name": "BaseBdev4", 00:10:37.215 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:37.215 "is_configured": true, 00:10:37.215 "data_offset": 2048, 00:10:37.215 "data_size": 63488 00:10:37.215 } 00:10:37.215 ] 00:10:37.215 }' 00:10:37.215 16:48:58 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.216 16:48:58 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.788 [2024-09-29 16:48:59.164864] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:37.788 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:37.789 "name": "Existed_Raid", 00:10:37.789 "aliases": [ 00:10:37.789 "e49467a3-bcac-4209-a7fa-7d1fe630dd02" 00:10:37.789 ], 00:10:37.789 "product_name": "Raid Volume", 00:10:37.789 "block_size": 512, 00:10:37.789 "num_blocks": 63488, 00:10:37.789 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:37.789 "assigned_rate_limits": { 00:10:37.789 "rw_ios_per_sec": 0, 00:10:37.789 "rw_mbytes_per_sec": 0, 00:10:37.789 "r_mbytes_per_sec": 0, 00:10:37.789 "w_mbytes_per_sec": 0 00:10:37.789 }, 00:10:37.789 "claimed": false, 00:10:37.789 "zoned": false, 00:10:37.789 "supported_io_types": { 00:10:37.789 "read": true, 00:10:37.789 "write": true, 00:10:37.789 "unmap": false, 00:10:37.789 "flush": false, 00:10:37.789 "reset": true, 00:10:37.789 "nvme_admin": false, 00:10:37.789 "nvme_io": false, 00:10:37.789 "nvme_io_md": false, 00:10:37.789 "write_zeroes": true, 00:10:37.789 "zcopy": false, 00:10:37.789 "get_zone_info": false, 00:10:37.789 "zone_management": false, 00:10:37.789 "zone_append": false, 00:10:37.789 "compare": false, 00:10:37.789 "compare_and_write": false, 00:10:37.789 "abort": false, 00:10:37.789 "seek_hole": false, 00:10:37.789 "seek_data": false, 00:10:37.789 "copy": false, 00:10:37.789 "nvme_iov_md": false 00:10:37.789 }, 00:10:37.789 "memory_domains": [ 00:10:37.789 { 00:10:37.789 "dma_device_id": "system", 00:10:37.789 "dma_device_type": 1 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.789 "dma_device_type": 2 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "dma_device_id": "system", 00:10:37.789 "dma_device_type": 1 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.789 "dma_device_type": 2 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "dma_device_id": "system", 00:10:37.789 "dma_device_type": 1 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.789 "dma_device_type": 2 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "dma_device_id": "system", 00:10:37.789 "dma_device_type": 1 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.789 "dma_device_type": 2 00:10:37.789 } 00:10:37.789 ], 00:10:37.789 "driver_specific": { 00:10:37.789 "raid": { 00:10:37.789 "uuid": "e49467a3-bcac-4209-a7fa-7d1fe630dd02", 00:10:37.789 "strip_size_kb": 0, 00:10:37.789 "state": "online", 00:10:37.789 "raid_level": "raid1", 00:10:37.789 "superblock": true, 00:10:37.789 "num_base_bdevs": 4, 00:10:37.789 "num_base_bdevs_discovered": 4, 00:10:37.789 "num_base_bdevs_operational": 4, 00:10:37.789 "base_bdevs_list": [ 00:10:37.789 { 00:10:37.789 "name": "NewBaseBdev", 00:10:37.789 "uuid": "0ea572f9-86f7-40de-a29a-f2eaf000875a", 00:10:37.789 "is_configured": true, 00:10:37.789 "data_offset": 2048, 00:10:37.789 "data_size": 63488 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "name": "BaseBdev2", 00:10:37.789 "uuid": "2db53c81-c920-4faf-aaea-eb5f86547e57", 00:10:37.789 "is_configured": true, 00:10:37.789 "data_offset": 2048, 00:10:37.789 "data_size": 63488 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "name": "BaseBdev3", 00:10:37.789 "uuid": "5059030f-c0c0-4bd9-b786-7851b22753d6", 00:10:37.789 "is_configured": true, 00:10:37.789 "data_offset": 2048, 00:10:37.789 "data_size": 63488 00:10:37.789 }, 00:10:37.789 { 00:10:37.789 "name": "BaseBdev4", 00:10:37.789 "uuid": "3c169108-45d7-4e85-b0a9-f819daab7d0f", 00:10:37.789 "is_configured": true, 00:10:37.789 "data_offset": 2048, 00:10:37.789 "data_size": 63488 00:10:37.789 } 00:10:37.789 ] 00:10:37.789 } 00:10:37.789 } 00:10:37.789 }' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:37.789 BaseBdev2 00:10:37.789 BaseBdev3 00:10:37.789 BaseBdev4' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:37.789 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.051 [2024-09-29 16:48:59.495973] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:38.051 [2024-09-29 16:48:59.495998] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:38.051 [2024-09-29 16:48:59.496067] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:38.051 [2024-09-29 16:48:59.496309] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:38.051 [2024-09-29 16:48:59.496323] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84303 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 84303 ']' 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 84303 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84303 00:10:38.051 killing process with pid 84303 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84303' 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 84303 00:10:38.051 [2024-09-29 16:48:59.538112] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:38.051 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 84303 00:10:38.051 [2024-09-29 16:48:59.578833] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:38.311 16:48:59 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:38.311 00:10:38.311 real 0m9.477s 00:10:38.311 user 0m16.228s 00:10:38.311 sys 0m1.887s 00:10:38.311 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:38.311 16:48:59 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.311 ************************************ 00:10:38.311 END TEST raid_state_function_test_sb 00:10:38.311 ************************************ 00:10:38.311 16:48:59 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:10:38.311 16:48:59 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:38.311 16:48:59 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:38.311 16:48:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:38.311 ************************************ 00:10:38.311 START TEST raid_superblock_test 00:10:38.311 ************************************ 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 4 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:10:38.311 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=84953 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 84953 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 84953 ']' 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:38.311 16:48:59 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.571 [2024-09-29 16:48:59.989958] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:38.571 [2024-09-29 16:48:59.990196] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid84953 ] 00:10:38.571 [2024-09-29 16:49:00.116585] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:38.571 [2024-09-29 16:49:00.165359] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:38.571 [2024-09-29 16:49:00.207337] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:38.571 [2024-09-29 16:49:00.207459] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.140 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.400 malloc1 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.400 [2024-09-29 16:49:00.833339] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:39.400 [2024-09-29 16:49:00.833460] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.400 [2024-09-29 16:49:00.833506] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:39.400 [2024-09-29 16:49:00.833539] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.400 [2024-09-29 16:49:00.835668] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.400 [2024-09-29 16:49:00.835758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:39.400 pt1 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:39.400 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 malloc2 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 [2024-09-29 16:49:00.882107] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:39.401 [2024-09-29 16:49:00.882242] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.401 [2024-09-29 16:49:00.882289] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:39.401 [2024-09-29 16:49:00.882321] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.401 [2024-09-29 16:49:00.886776] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.401 [2024-09-29 16:49:00.886841] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:39.401 pt2 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 malloc3 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 [2024-09-29 16:49:00.912568] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:39.401 [2024-09-29 16:49:00.912663] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.401 [2024-09-29 16:49:00.912698] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:39.401 [2024-09-29 16:49:00.912745] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.401 [2024-09-29 16:49:00.914777] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.401 [2024-09-29 16:49:00.914862] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:39.401 pt3 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 malloc4 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 [2024-09-29 16:49:00.945219] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:39.401 [2024-09-29 16:49:00.945312] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:39.401 [2024-09-29 16:49:00.945344] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:39.401 [2024-09-29 16:49:00.945375] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:39.401 [2024-09-29 16:49:00.947453] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:39.401 [2024-09-29 16:49:00.947527] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:39.401 pt4 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 [2024-09-29 16:49:00.957254] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:39.401 [2024-09-29 16:49:00.959083] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:39.401 [2024-09-29 16:49:00.959152] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:39.401 [2024-09-29 16:49:00.959195] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:39.401 [2024-09-29 16:49:00.959352] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:39.401 [2024-09-29 16:49:00.959365] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:39.401 [2024-09-29 16:49:00.959592] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:39.401 [2024-09-29 16:49:00.959745] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:39.401 [2024-09-29 16:49:00.959755] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:39.401 [2024-09-29 16:49:00.959891] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.401 16:49:00 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.401 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.401 "name": "raid_bdev1", 00:10:39.401 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:39.401 "strip_size_kb": 0, 00:10:39.401 "state": "online", 00:10:39.401 "raid_level": "raid1", 00:10:39.401 "superblock": true, 00:10:39.401 "num_base_bdevs": 4, 00:10:39.401 "num_base_bdevs_discovered": 4, 00:10:39.401 "num_base_bdevs_operational": 4, 00:10:39.401 "base_bdevs_list": [ 00:10:39.401 { 00:10:39.401 "name": "pt1", 00:10:39.401 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:39.401 "is_configured": true, 00:10:39.401 "data_offset": 2048, 00:10:39.401 "data_size": 63488 00:10:39.401 }, 00:10:39.401 { 00:10:39.401 "name": "pt2", 00:10:39.401 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:39.401 "is_configured": true, 00:10:39.401 "data_offset": 2048, 00:10:39.401 "data_size": 63488 00:10:39.401 }, 00:10:39.401 { 00:10:39.401 "name": "pt3", 00:10:39.401 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:39.401 "is_configured": true, 00:10:39.401 "data_offset": 2048, 00:10:39.401 "data_size": 63488 00:10:39.401 }, 00:10:39.401 { 00:10:39.401 "name": "pt4", 00:10:39.402 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:39.402 "is_configured": true, 00:10:39.402 "data_offset": 2048, 00:10:39.402 "data_size": 63488 00:10:39.402 } 00:10:39.402 ] 00:10:39.402 }' 00:10:39.402 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.402 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.970 [2024-09-29 16:49:01.364854] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.970 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:39.970 "name": "raid_bdev1", 00:10:39.970 "aliases": [ 00:10:39.970 "4a070d8d-32f8-46ad-afe2-62ce91acae2d" 00:10:39.970 ], 00:10:39.970 "product_name": "Raid Volume", 00:10:39.970 "block_size": 512, 00:10:39.970 "num_blocks": 63488, 00:10:39.970 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:39.970 "assigned_rate_limits": { 00:10:39.970 "rw_ios_per_sec": 0, 00:10:39.970 "rw_mbytes_per_sec": 0, 00:10:39.970 "r_mbytes_per_sec": 0, 00:10:39.970 "w_mbytes_per_sec": 0 00:10:39.970 }, 00:10:39.970 "claimed": false, 00:10:39.970 "zoned": false, 00:10:39.970 "supported_io_types": { 00:10:39.970 "read": true, 00:10:39.970 "write": true, 00:10:39.971 "unmap": false, 00:10:39.971 "flush": false, 00:10:39.971 "reset": true, 00:10:39.971 "nvme_admin": false, 00:10:39.971 "nvme_io": false, 00:10:39.971 "nvme_io_md": false, 00:10:39.971 "write_zeroes": true, 00:10:39.971 "zcopy": false, 00:10:39.971 "get_zone_info": false, 00:10:39.971 "zone_management": false, 00:10:39.971 "zone_append": false, 00:10:39.971 "compare": false, 00:10:39.971 "compare_and_write": false, 00:10:39.971 "abort": false, 00:10:39.971 "seek_hole": false, 00:10:39.971 "seek_data": false, 00:10:39.971 "copy": false, 00:10:39.971 "nvme_iov_md": false 00:10:39.971 }, 00:10:39.971 "memory_domains": [ 00:10:39.971 { 00:10:39.971 "dma_device_id": "system", 00:10:39.971 "dma_device_type": 1 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.971 "dma_device_type": 2 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "dma_device_id": "system", 00:10:39.971 "dma_device_type": 1 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.971 "dma_device_type": 2 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "dma_device_id": "system", 00:10:39.971 "dma_device_type": 1 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.971 "dma_device_type": 2 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "dma_device_id": "system", 00:10:39.971 "dma_device_type": 1 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.971 "dma_device_type": 2 00:10:39.971 } 00:10:39.971 ], 00:10:39.971 "driver_specific": { 00:10:39.971 "raid": { 00:10:39.971 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:39.971 "strip_size_kb": 0, 00:10:39.971 "state": "online", 00:10:39.971 "raid_level": "raid1", 00:10:39.971 "superblock": true, 00:10:39.971 "num_base_bdevs": 4, 00:10:39.971 "num_base_bdevs_discovered": 4, 00:10:39.971 "num_base_bdevs_operational": 4, 00:10:39.971 "base_bdevs_list": [ 00:10:39.971 { 00:10:39.971 "name": "pt1", 00:10:39.971 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:39.971 "is_configured": true, 00:10:39.971 "data_offset": 2048, 00:10:39.971 "data_size": 63488 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "name": "pt2", 00:10:39.971 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:39.971 "is_configured": true, 00:10:39.971 "data_offset": 2048, 00:10:39.971 "data_size": 63488 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "name": "pt3", 00:10:39.971 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:39.971 "is_configured": true, 00:10:39.971 "data_offset": 2048, 00:10:39.971 "data_size": 63488 00:10:39.971 }, 00:10:39.971 { 00:10:39.971 "name": "pt4", 00:10:39.971 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:39.971 "is_configured": true, 00:10:39.971 "data_offset": 2048, 00:10:39.971 "data_size": 63488 00:10:39.971 } 00:10:39.971 ] 00:10:39.971 } 00:10:39.971 } 00:10:39.971 }' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:39.971 pt2 00:10:39.971 pt3 00:10:39.971 pt4' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.971 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 [2024-09-29 16:49:01.676243] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=4a070d8d-32f8-46ad-afe2-62ce91acae2d 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 4a070d8d-32f8-46ad-afe2-62ce91acae2d ']' 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 [2024-09-29 16:49:01.723884] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:40.231 [2024-09-29 16:49:01.723952] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:40.231 [2024-09-29 16:49:01.724035] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:40.231 [2024-09-29 16:49:01.724162] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:40.231 [2024-09-29 16:49:01.724172] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.231 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.231 [2024-09-29 16:49:01.895636] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:40.231 [2024-09-29 16:49:01.897476] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:40.231 [2024-09-29 16:49:01.897522] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:40.231 [2024-09-29 16:49:01.897555] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:40.231 [2024-09-29 16:49:01.897602] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:40.231 [2024-09-29 16:49:01.897652] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:40.231 [2024-09-29 16:49:01.897673] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:40.231 [2024-09-29 16:49:01.897687] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:40.231 [2024-09-29 16:49:01.897701] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:40.231 [2024-09-29 16:49:01.897710] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:40.491 request: 00:10:40.491 { 00:10:40.491 "name": "raid_bdev1", 00:10:40.491 "raid_level": "raid1", 00:10:40.491 "base_bdevs": [ 00:10:40.491 "malloc1", 00:10:40.491 "malloc2", 00:10:40.491 "malloc3", 00:10:40.491 "malloc4" 00:10:40.491 ], 00:10:40.491 "superblock": false, 00:10:40.491 "method": "bdev_raid_create", 00:10:40.491 "req_id": 1 00:10:40.491 } 00:10:40.491 Got JSON-RPC error response 00:10:40.491 response: 00:10:40.491 { 00:10:40.491 "code": -17, 00:10:40.491 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:40.491 } 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.491 [2024-09-29 16:49:01.935528] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:40.491 [2024-09-29 16:49:01.935629] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:40.491 [2024-09-29 16:49:01.935665] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:40.491 [2024-09-29 16:49:01.935693] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:40.491 [2024-09-29 16:49:01.937820] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:40.491 [2024-09-29 16:49:01.937885] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:40.491 [2024-09-29 16:49:01.937972] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:40.491 [2024-09-29 16:49:01.938066] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:40.491 pt1 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.491 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.492 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.492 "name": "raid_bdev1", 00:10:40.492 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:40.492 "strip_size_kb": 0, 00:10:40.492 "state": "configuring", 00:10:40.492 "raid_level": "raid1", 00:10:40.492 "superblock": true, 00:10:40.492 "num_base_bdevs": 4, 00:10:40.492 "num_base_bdevs_discovered": 1, 00:10:40.492 "num_base_bdevs_operational": 4, 00:10:40.492 "base_bdevs_list": [ 00:10:40.492 { 00:10:40.492 "name": "pt1", 00:10:40.492 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:40.492 "is_configured": true, 00:10:40.492 "data_offset": 2048, 00:10:40.492 "data_size": 63488 00:10:40.492 }, 00:10:40.492 { 00:10:40.492 "name": null, 00:10:40.492 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:40.492 "is_configured": false, 00:10:40.492 "data_offset": 2048, 00:10:40.492 "data_size": 63488 00:10:40.492 }, 00:10:40.492 { 00:10:40.492 "name": null, 00:10:40.492 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:40.492 "is_configured": false, 00:10:40.492 "data_offset": 2048, 00:10:40.492 "data_size": 63488 00:10:40.492 }, 00:10:40.492 { 00:10:40.492 "name": null, 00:10:40.492 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:40.492 "is_configured": false, 00:10:40.492 "data_offset": 2048, 00:10:40.492 "data_size": 63488 00:10:40.492 } 00:10:40.492 ] 00:10:40.492 }' 00:10:40.492 16:49:01 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.492 16:49:01 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.751 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:40.751 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.752 [2024-09-29 16:49:02.347002] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:40.752 [2024-09-29 16:49:02.347096] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:40.752 [2024-09-29 16:49:02.347120] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:40.752 [2024-09-29 16:49:02.347129] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:40.752 [2024-09-29 16:49:02.347538] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:40.752 [2024-09-29 16:49:02.347556] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:40.752 [2024-09-29 16:49:02.347627] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:40.752 [2024-09-29 16:49:02.347658] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:40.752 pt2 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.752 [2024-09-29 16:49:02.355046] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.752 "name": "raid_bdev1", 00:10:40.752 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:40.752 "strip_size_kb": 0, 00:10:40.752 "state": "configuring", 00:10:40.752 "raid_level": "raid1", 00:10:40.752 "superblock": true, 00:10:40.752 "num_base_bdevs": 4, 00:10:40.752 "num_base_bdevs_discovered": 1, 00:10:40.752 "num_base_bdevs_operational": 4, 00:10:40.752 "base_bdevs_list": [ 00:10:40.752 { 00:10:40.752 "name": "pt1", 00:10:40.752 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:40.752 "is_configured": true, 00:10:40.752 "data_offset": 2048, 00:10:40.752 "data_size": 63488 00:10:40.752 }, 00:10:40.752 { 00:10:40.752 "name": null, 00:10:40.752 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:40.752 "is_configured": false, 00:10:40.752 "data_offset": 0, 00:10:40.752 "data_size": 63488 00:10:40.752 }, 00:10:40.752 { 00:10:40.752 "name": null, 00:10:40.752 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:40.752 "is_configured": false, 00:10:40.752 "data_offset": 2048, 00:10:40.752 "data_size": 63488 00:10:40.752 }, 00:10:40.752 { 00:10:40.752 "name": null, 00:10:40.752 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:40.752 "is_configured": false, 00:10:40.752 "data_offset": 2048, 00:10:40.752 "data_size": 63488 00:10:40.752 } 00:10:40.752 ] 00:10:40.752 }' 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.752 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.321 [2024-09-29 16:49:02.818320] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:41.321 [2024-09-29 16:49:02.818419] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.321 [2024-09-29 16:49:02.818439] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:41.321 [2024-09-29 16:49:02.818449] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.321 [2024-09-29 16:49:02.818835] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.321 [2024-09-29 16:49:02.818862] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:41.321 [2024-09-29 16:49:02.818954] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:41.321 [2024-09-29 16:49:02.818978] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:41.321 pt2 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.321 [2024-09-29 16:49:02.830280] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:41.321 [2024-09-29 16:49:02.830325] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.321 [2024-09-29 16:49:02.830348] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:41.321 [2024-09-29 16:49:02.830358] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.321 [2024-09-29 16:49:02.830662] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.321 [2024-09-29 16:49:02.830679] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:41.321 [2024-09-29 16:49:02.830750] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:41.321 [2024-09-29 16:49:02.830772] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:41.321 pt3 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.321 [2024-09-29 16:49:02.842265] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:41.321 [2024-09-29 16:49:02.842314] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:41.321 [2024-09-29 16:49:02.842328] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:41.321 [2024-09-29 16:49:02.842336] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:41.321 [2024-09-29 16:49:02.842596] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:41.321 [2024-09-29 16:49:02.842613] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:41.321 [2024-09-29 16:49:02.842657] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:41.321 [2024-09-29 16:49:02.842674] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:41.321 [2024-09-29 16:49:02.842801] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:41.321 [2024-09-29 16:49:02.842812] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:41.321 [2024-09-29 16:49:02.843045] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:41.321 [2024-09-29 16:49:02.843167] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:41.321 [2024-09-29 16:49:02.843182] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:41.321 [2024-09-29 16:49:02.843330] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:41.321 pt4 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.321 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.321 "name": "raid_bdev1", 00:10:41.321 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:41.321 "strip_size_kb": 0, 00:10:41.321 "state": "online", 00:10:41.321 "raid_level": "raid1", 00:10:41.321 "superblock": true, 00:10:41.321 "num_base_bdevs": 4, 00:10:41.321 "num_base_bdevs_discovered": 4, 00:10:41.321 "num_base_bdevs_operational": 4, 00:10:41.321 "base_bdevs_list": [ 00:10:41.321 { 00:10:41.321 "name": "pt1", 00:10:41.321 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:41.321 "is_configured": true, 00:10:41.321 "data_offset": 2048, 00:10:41.321 "data_size": 63488 00:10:41.321 }, 00:10:41.321 { 00:10:41.321 "name": "pt2", 00:10:41.321 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:41.321 "is_configured": true, 00:10:41.321 "data_offset": 2048, 00:10:41.321 "data_size": 63488 00:10:41.321 }, 00:10:41.321 { 00:10:41.321 "name": "pt3", 00:10:41.321 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:41.321 "is_configured": true, 00:10:41.321 "data_offset": 2048, 00:10:41.321 "data_size": 63488 00:10:41.321 }, 00:10:41.321 { 00:10:41.321 "name": "pt4", 00:10:41.321 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:41.321 "is_configured": true, 00:10:41.321 "data_offset": 2048, 00:10:41.321 "data_size": 63488 00:10:41.321 } 00:10:41.321 ] 00:10:41.321 }' 00:10:41.322 16:49:02 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.322 16:49:02 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.890 [2024-09-29 16:49:03.293810] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.890 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:41.890 "name": "raid_bdev1", 00:10:41.890 "aliases": [ 00:10:41.890 "4a070d8d-32f8-46ad-afe2-62ce91acae2d" 00:10:41.890 ], 00:10:41.890 "product_name": "Raid Volume", 00:10:41.890 "block_size": 512, 00:10:41.890 "num_blocks": 63488, 00:10:41.890 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:41.890 "assigned_rate_limits": { 00:10:41.890 "rw_ios_per_sec": 0, 00:10:41.890 "rw_mbytes_per_sec": 0, 00:10:41.890 "r_mbytes_per_sec": 0, 00:10:41.890 "w_mbytes_per_sec": 0 00:10:41.890 }, 00:10:41.890 "claimed": false, 00:10:41.890 "zoned": false, 00:10:41.890 "supported_io_types": { 00:10:41.890 "read": true, 00:10:41.890 "write": true, 00:10:41.890 "unmap": false, 00:10:41.890 "flush": false, 00:10:41.890 "reset": true, 00:10:41.890 "nvme_admin": false, 00:10:41.890 "nvme_io": false, 00:10:41.891 "nvme_io_md": false, 00:10:41.891 "write_zeroes": true, 00:10:41.891 "zcopy": false, 00:10:41.891 "get_zone_info": false, 00:10:41.891 "zone_management": false, 00:10:41.891 "zone_append": false, 00:10:41.891 "compare": false, 00:10:41.891 "compare_and_write": false, 00:10:41.891 "abort": false, 00:10:41.891 "seek_hole": false, 00:10:41.891 "seek_data": false, 00:10:41.891 "copy": false, 00:10:41.891 "nvme_iov_md": false 00:10:41.891 }, 00:10:41.891 "memory_domains": [ 00:10:41.891 { 00:10:41.891 "dma_device_id": "system", 00:10:41.891 "dma_device_type": 1 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.891 "dma_device_type": 2 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "dma_device_id": "system", 00:10:41.891 "dma_device_type": 1 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.891 "dma_device_type": 2 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "dma_device_id": "system", 00:10:41.891 "dma_device_type": 1 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.891 "dma_device_type": 2 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "dma_device_id": "system", 00:10:41.891 "dma_device_type": 1 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.891 "dma_device_type": 2 00:10:41.891 } 00:10:41.891 ], 00:10:41.891 "driver_specific": { 00:10:41.891 "raid": { 00:10:41.891 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:41.891 "strip_size_kb": 0, 00:10:41.891 "state": "online", 00:10:41.891 "raid_level": "raid1", 00:10:41.891 "superblock": true, 00:10:41.891 "num_base_bdevs": 4, 00:10:41.891 "num_base_bdevs_discovered": 4, 00:10:41.891 "num_base_bdevs_operational": 4, 00:10:41.891 "base_bdevs_list": [ 00:10:41.891 { 00:10:41.891 "name": "pt1", 00:10:41.891 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:41.891 "is_configured": true, 00:10:41.891 "data_offset": 2048, 00:10:41.891 "data_size": 63488 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "name": "pt2", 00:10:41.891 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:41.891 "is_configured": true, 00:10:41.891 "data_offset": 2048, 00:10:41.891 "data_size": 63488 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "name": "pt3", 00:10:41.891 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:41.891 "is_configured": true, 00:10:41.891 "data_offset": 2048, 00:10:41.891 "data_size": 63488 00:10:41.891 }, 00:10:41.891 { 00:10:41.891 "name": "pt4", 00:10:41.891 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:41.891 "is_configured": true, 00:10:41.891 "data_offset": 2048, 00:10:41.891 "data_size": 63488 00:10:41.891 } 00:10:41.891 ] 00:10:41.891 } 00:10:41.891 } 00:10:41.891 }' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:41.891 pt2 00:10:41.891 pt3 00:10:41.891 pt4' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:41.891 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:42.150 [2024-09-29 16:49:03.625177] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 4a070d8d-32f8-46ad-afe2-62ce91acae2d '!=' 4a070d8d-32f8-46ad-afe2-62ce91acae2d ']' 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.150 [2024-09-29 16:49:03.668855] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.150 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.150 "name": "raid_bdev1", 00:10:42.150 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:42.150 "strip_size_kb": 0, 00:10:42.150 "state": "online", 00:10:42.150 "raid_level": "raid1", 00:10:42.150 "superblock": true, 00:10:42.150 "num_base_bdevs": 4, 00:10:42.150 "num_base_bdevs_discovered": 3, 00:10:42.150 "num_base_bdevs_operational": 3, 00:10:42.150 "base_bdevs_list": [ 00:10:42.150 { 00:10:42.150 "name": null, 00:10:42.150 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.150 "is_configured": false, 00:10:42.150 "data_offset": 0, 00:10:42.150 "data_size": 63488 00:10:42.150 }, 00:10:42.150 { 00:10:42.150 "name": "pt2", 00:10:42.150 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:42.150 "is_configured": true, 00:10:42.150 "data_offset": 2048, 00:10:42.150 "data_size": 63488 00:10:42.150 }, 00:10:42.150 { 00:10:42.150 "name": "pt3", 00:10:42.150 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:42.150 "is_configured": true, 00:10:42.150 "data_offset": 2048, 00:10:42.150 "data_size": 63488 00:10:42.150 }, 00:10:42.150 { 00:10:42.150 "name": "pt4", 00:10:42.151 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:42.151 "is_configured": true, 00:10:42.151 "data_offset": 2048, 00:10:42.151 "data_size": 63488 00:10:42.151 } 00:10:42.151 ] 00:10:42.151 }' 00:10:42.151 16:49:03 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.151 16:49:03 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.719 [2024-09-29 16:49:04.096090] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:42.719 [2024-09-29 16:49:04.096164] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:42.719 [2024-09-29 16:49:04.096271] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:42.719 [2024-09-29 16:49:04.096384] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:42.719 [2024-09-29 16:49:04.096434] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.719 [2024-09-29 16:49:04.175935] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:42.719 [2024-09-29 16:49:04.175986] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.719 [2024-09-29 16:49:04.176001] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:10:42.719 [2024-09-29 16:49:04.176011] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.719 [2024-09-29 16:49:04.178104] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.719 [2024-09-29 16:49:04.178142] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:42.719 [2024-09-29 16:49:04.178210] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:42.719 [2024-09-29 16:49:04.178254] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:42.719 pt2 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.719 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.720 "name": "raid_bdev1", 00:10:42.720 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:42.720 "strip_size_kb": 0, 00:10:42.720 "state": "configuring", 00:10:42.720 "raid_level": "raid1", 00:10:42.720 "superblock": true, 00:10:42.720 "num_base_bdevs": 4, 00:10:42.720 "num_base_bdevs_discovered": 1, 00:10:42.720 "num_base_bdevs_operational": 3, 00:10:42.720 "base_bdevs_list": [ 00:10:42.720 { 00:10:42.720 "name": null, 00:10:42.720 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.720 "is_configured": false, 00:10:42.720 "data_offset": 2048, 00:10:42.720 "data_size": 63488 00:10:42.720 }, 00:10:42.720 { 00:10:42.720 "name": "pt2", 00:10:42.720 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:42.720 "is_configured": true, 00:10:42.720 "data_offset": 2048, 00:10:42.720 "data_size": 63488 00:10:42.720 }, 00:10:42.720 { 00:10:42.720 "name": null, 00:10:42.720 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:42.720 "is_configured": false, 00:10:42.720 "data_offset": 2048, 00:10:42.720 "data_size": 63488 00:10:42.720 }, 00:10:42.720 { 00:10:42.720 "name": null, 00:10:42.720 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:42.720 "is_configured": false, 00:10:42.720 "data_offset": 2048, 00:10:42.720 "data_size": 63488 00:10:42.720 } 00:10:42.720 ] 00:10:42.720 }' 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.720 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.979 [2024-09-29 16:49:04.615211] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:42.979 [2024-09-29 16:49:04.615307] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:42.979 [2024-09-29 16:49:04.615342] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:10:42.979 [2024-09-29 16:49:04.615372] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:42.979 [2024-09-29 16:49:04.615847] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:42.979 [2024-09-29 16:49:04.615915] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:42.979 [2024-09-29 16:49:04.616036] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:42.979 [2024-09-29 16:49:04.616117] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:42.979 pt3 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:42.979 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.239 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.239 "name": "raid_bdev1", 00:10:43.239 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:43.239 "strip_size_kb": 0, 00:10:43.239 "state": "configuring", 00:10:43.239 "raid_level": "raid1", 00:10:43.239 "superblock": true, 00:10:43.239 "num_base_bdevs": 4, 00:10:43.239 "num_base_bdevs_discovered": 2, 00:10:43.239 "num_base_bdevs_operational": 3, 00:10:43.239 "base_bdevs_list": [ 00:10:43.239 { 00:10:43.239 "name": null, 00:10:43.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.239 "is_configured": false, 00:10:43.239 "data_offset": 2048, 00:10:43.239 "data_size": 63488 00:10:43.239 }, 00:10:43.239 { 00:10:43.239 "name": "pt2", 00:10:43.239 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:43.239 "is_configured": true, 00:10:43.239 "data_offset": 2048, 00:10:43.239 "data_size": 63488 00:10:43.239 }, 00:10:43.239 { 00:10:43.239 "name": "pt3", 00:10:43.239 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:43.239 "is_configured": true, 00:10:43.239 "data_offset": 2048, 00:10:43.239 "data_size": 63488 00:10:43.239 }, 00:10:43.239 { 00:10:43.239 "name": null, 00:10:43.239 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:43.239 "is_configured": false, 00:10:43.239 "data_offset": 2048, 00:10:43.239 "data_size": 63488 00:10:43.239 } 00:10:43.239 ] 00:10:43.239 }' 00:10:43.239 16:49:04 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.239 16:49:04 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.502 [2024-09-29 16:49:05.042513] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:43.502 [2024-09-29 16:49:05.042570] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:43.502 [2024-09-29 16:49:05.042588] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:10:43.502 [2024-09-29 16:49:05.042598] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:43.502 [2024-09-29 16:49:05.043027] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:43.502 [2024-09-29 16:49:05.043055] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:43.502 [2024-09-29 16:49:05.043127] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:43.502 [2024-09-29 16:49:05.043150] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:43.502 [2024-09-29 16:49:05.043247] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:43.502 [2024-09-29 16:49:05.043264] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:43.502 [2024-09-29 16:49:05.043493] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:43.502 [2024-09-29 16:49:05.043618] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:43.502 [2024-09-29 16:49:05.043627] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:10:43.502 [2024-09-29 16:49:05.043786] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:43.502 pt4 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.502 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.502 "name": "raid_bdev1", 00:10:43.502 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:43.502 "strip_size_kb": 0, 00:10:43.502 "state": "online", 00:10:43.502 "raid_level": "raid1", 00:10:43.502 "superblock": true, 00:10:43.502 "num_base_bdevs": 4, 00:10:43.503 "num_base_bdevs_discovered": 3, 00:10:43.503 "num_base_bdevs_operational": 3, 00:10:43.503 "base_bdevs_list": [ 00:10:43.503 { 00:10:43.503 "name": null, 00:10:43.503 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.503 "is_configured": false, 00:10:43.503 "data_offset": 2048, 00:10:43.503 "data_size": 63488 00:10:43.503 }, 00:10:43.503 { 00:10:43.503 "name": "pt2", 00:10:43.503 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:43.503 "is_configured": true, 00:10:43.503 "data_offset": 2048, 00:10:43.503 "data_size": 63488 00:10:43.503 }, 00:10:43.503 { 00:10:43.503 "name": "pt3", 00:10:43.503 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:43.503 "is_configured": true, 00:10:43.503 "data_offset": 2048, 00:10:43.503 "data_size": 63488 00:10:43.503 }, 00:10:43.503 { 00:10:43.503 "name": "pt4", 00:10:43.503 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:43.503 "is_configured": true, 00:10:43.503 "data_offset": 2048, 00:10:43.503 "data_size": 63488 00:10:43.503 } 00:10:43.503 ] 00:10:43.503 }' 00:10:43.503 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.503 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.825 [2024-09-29 16:49:05.417880] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:43.825 [2024-09-29 16:49:05.417954] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:43.825 [2024-09-29 16:49:05.418041] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:43.825 [2024-09-29 16:49:05.418165] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:43.825 [2024-09-29 16:49:05.418213] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:43.825 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.084 [2024-09-29 16:49:05.489734] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:44.084 [2024-09-29 16:49:05.489832] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:44.084 [2024-09-29 16:49:05.489872] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:10:44.084 [2024-09-29 16:49:05.489900] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:44.084 [2024-09-29 16:49:05.492136] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:44.084 [2024-09-29 16:49:05.492222] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:44.084 [2024-09-29 16:49:05.492345] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:44.084 [2024-09-29 16:49:05.492429] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:44.084 [2024-09-29 16:49:05.492601] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:10:44.084 [2024-09-29 16:49:05.492660] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:44.084 [2024-09-29 16:49:05.492717] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:10:44.084 [2024-09-29 16:49:05.492818] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:44.084 [2024-09-29 16:49:05.492989] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:44.084 pt1 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.084 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.084 "name": "raid_bdev1", 00:10:44.084 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:44.084 "strip_size_kb": 0, 00:10:44.084 "state": "configuring", 00:10:44.084 "raid_level": "raid1", 00:10:44.084 "superblock": true, 00:10:44.084 "num_base_bdevs": 4, 00:10:44.084 "num_base_bdevs_discovered": 2, 00:10:44.084 "num_base_bdevs_operational": 3, 00:10:44.084 "base_bdevs_list": [ 00:10:44.084 { 00:10:44.084 "name": null, 00:10:44.084 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.084 "is_configured": false, 00:10:44.084 "data_offset": 2048, 00:10:44.084 "data_size": 63488 00:10:44.085 }, 00:10:44.085 { 00:10:44.085 "name": "pt2", 00:10:44.085 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:44.085 "is_configured": true, 00:10:44.085 "data_offset": 2048, 00:10:44.085 "data_size": 63488 00:10:44.085 }, 00:10:44.085 { 00:10:44.085 "name": "pt3", 00:10:44.085 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:44.085 "is_configured": true, 00:10:44.085 "data_offset": 2048, 00:10:44.085 "data_size": 63488 00:10:44.085 }, 00:10:44.085 { 00:10:44.085 "name": null, 00:10:44.085 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:44.085 "is_configured": false, 00:10:44.085 "data_offset": 2048, 00:10:44.085 "data_size": 63488 00:10:44.085 } 00:10:44.085 ] 00:10:44.085 }' 00:10:44.085 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.085 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.343 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:10:44.343 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.343 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.343 16:49:05 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:44.343 16:49:05 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.343 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:10:44.343 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:44.343 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.343 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.343 [2024-09-29 16:49:06.012851] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:44.343 [2024-09-29 16:49:06.012907] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:44.343 [2024-09-29 16:49:06.012925] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:10:44.343 [2024-09-29 16:49:06.012935] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:44.343 [2024-09-29 16:49:06.013308] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:44.343 [2024-09-29 16:49:06.013340] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:44.343 [2024-09-29 16:49:06.013407] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:44.343 [2024-09-29 16:49:06.013431] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:44.343 [2024-09-29 16:49:06.013561] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:10:44.343 [2024-09-29 16:49:06.013580] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:44.343 [2024-09-29 16:49:06.013835] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:10:44.343 [2024-09-29 16:49:06.013966] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:10:44.343 [2024-09-29 16:49:06.013976] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:10:44.343 [2024-09-29 16:49:06.014108] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:44.601 pt4 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.601 "name": "raid_bdev1", 00:10:44.601 "uuid": "4a070d8d-32f8-46ad-afe2-62ce91acae2d", 00:10:44.601 "strip_size_kb": 0, 00:10:44.601 "state": "online", 00:10:44.601 "raid_level": "raid1", 00:10:44.601 "superblock": true, 00:10:44.601 "num_base_bdevs": 4, 00:10:44.601 "num_base_bdevs_discovered": 3, 00:10:44.601 "num_base_bdevs_operational": 3, 00:10:44.601 "base_bdevs_list": [ 00:10:44.601 { 00:10:44.601 "name": null, 00:10:44.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:44.601 "is_configured": false, 00:10:44.601 "data_offset": 2048, 00:10:44.601 "data_size": 63488 00:10:44.601 }, 00:10:44.601 { 00:10:44.601 "name": "pt2", 00:10:44.601 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:44.601 "is_configured": true, 00:10:44.601 "data_offset": 2048, 00:10:44.601 "data_size": 63488 00:10:44.601 }, 00:10:44.601 { 00:10:44.601 "name": "pt3", 00:10:44.601 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:44.601 "is_configured": true, 00:10:44.601 "data_offset": 2048, 00:10:44.601 "data_size": 63488 00:10:44.601 }, 00:10:44.601 { 00:10:44.601 "name": "pt4", 00:10:44.601 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:44.601 "is_configured": true, 00:10:44.601 "data_offset": 2048, 00:10:44.601 "data_size": 63488 00:10:44.601 } 00:10:44.601 ] 00:10:44.601 }' 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.601 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.859 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:10:44.859 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.859 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:44.859 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:10:44.859 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.195 [2024-09-29 16:49:06.544190] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 4a070d8d-32f8-46ad-afe2-62ce91acae2d '!=' 4a070d8d-32f8-46ad-afe2-62ce91acae2d ']' 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 84953 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 84953 ']' 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 84953 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84953 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84953' 00:10:45.195 killing process with pid 84953 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 84953 00:10:45.195 [2024-09-29 16:49:06.622830] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:45.195 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 84953 00:10:45.195 [2024-09-29 16:49:06.622983] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:45.195 [2024-09-29 16:49:06.623104] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:45.195 [2024-09-29 16:49:06.623151] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:10:45.195 [2024-09-29 16:49:06.667243] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:45.455 16:49:06 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:45.455 00:10:45.455 real 0m7.003s 00:10:45.455 user 0m11.763s 00:10:45.455 sys 0m1.481s 00:10:45.456 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:45.456 16:49:06 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.456 ************************************ 00:10:45.456 END TEST raid_superblock_test 00:10:45.456 ************************************ 00:10:45.456 16:49:06 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:10:45.456 16:49:06 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:45.456 16:49:06 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:45.456 16:49:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:45.456 ************************************ 00:10:45.456 START TEST raid_read_error_test 00:10:45.456 ************************************ 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 read 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.5J1Lz8YAhz 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85423 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85423 00:10:45.456 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 85423 ']' 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:45.456 16:49:06 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:45.456 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:45.456 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:45.456 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:45.456 [2024-09-29 16:49:07.076954] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:45.456 [2024-09-29 16:49:07.077068] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85423 ] 00:10:45.715 [2024-09-29 16:49:07.222623] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:45.715 [2024-09-29 16:49:07.268156] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:45.715 [2024-09-29 16:49:07.310057] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:45.715 [2024-09-29 16:49:07.310090] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.284 BaseBdev1_malloc 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.284 true 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.284 [2024-09-29 16:49:07.927987] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:46.284 [2024-09-29 16:49:07.928046] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:46.284 [2024-09-29 16:49:07.928067] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:46.284 [2024-09-29 16:49:07.928075] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:46.284 [2024-09-29 16:49:07.930206] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:46.284 [2024-09-29 16:49:07.930244] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:46.284 BaseBdev1 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.284 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.545 BaseBdev2_malloc 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.545 true 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.545 [2024-09-29 16:49:07.984332] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:46.545 [2024-09-29 16:49:07.984412] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:46.545 [2024-09-29 16:49:07.984443] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:46.545 [2024-09-29 16:49:07.984456] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:46.545 [2024-09-29 16:49:07.987324] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:46.545 [2024-09-29 16:49:07.987363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:46.545 BaseBdev2 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.545 16:49:07 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.545 BaseBdev3_malloc 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.545 true 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.545 [2024-09-29 16:49:08.024957] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:46.545 [2024-09-29 16:49:08.025012] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:46.545 [2024-09-29 16:49:08.025029] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:46.545 [2024-09-29 16:49:08.025037] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:46.545 [2024-09-29 16:49:08.027019] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:46.545 [2024-09-29 16:49:08.027054] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:46.545 BaseBdev3 00:10:46.545 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.546 BaseBdev4_malloc 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.546 true 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.546 [2024-09-29 16:49:08.065425] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:46.546 [2024-09-29 16:49:08.065470] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:46.546 [2024-09-29 16:49:08.065490] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:46.546 [2024-09-29 16:49:08.065498] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:46.546 [2024-09-29 16:49:08.067510] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:46.546 [2024-09-29 16:49:08.067546] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:46.546 BaseBdev4 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.546 [2024-09-29 16:49:08.077464] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:46.546 [2024-09-29 16:49:08.079298] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:46.546 [2024-09-29 16:49:08.079369] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:46.546 [2024-09-29 16:49:08.079430] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:46.546 [2024-09-29 16:49:08.079617] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:46.546 [2024-09-29 16:49:08.079628] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:46.546 [2024-09-29 16:49:08.079906] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:46.546 [2024-09-29 16:49:08.080060] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:46.546 [2024-09-29 16:49:08.080079] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:46.546 [2024-09-29 16:49:08.080216] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.546 "name": "raid_bdev1", 00:10:46.546 "uuid": "41bf150a-d3e1-42af-8420-0218d6725ca3", 00:10:46.546 "strip_size_kb": 0, 00:10:46.546 "state": "online", 00:10:46.546 "raid_level": "raid1", 00:10:46.546 "superblock": true, 00:10:46.546 "num_base_bdevs": 4, 00:10:46.546 "num_base_bdevs_discovered": 4, 00:10:46.546 "num_base_bdevs_operational": 4, 00:10:46.546 "base_bdevs_list": [ 00:10:46.546 { 00:10:46.546 "name": "BaseBdev1", 00:10:46.546 "uuid": "cbb3859b-06fd-5a76-8020-740ec659cefc", 00:10:46.546 "is_configured": true, 00:10:46.546 "data_offset": 2048, 00:10:46.546 "data_size": 63488 00:10:46.546 }, 00:10:46.546 { 00:10:46.546 "name": "BaseBdev2", 00:10:46.546 "uuid": "a6c9bf49-cd6d-533f-bda5-fe62fccb4681", 00:10:46.546 "is_configured": true, 00:10:46.546 "data_offset": 2048, 00:10:46.546 "data_size": 63488 00:10:46.546 }, 00:10:46.546 { 00:10:46.546 "name": "BaseBdev3", 00:10:46.546 "uuid": "c35c6a5a-ea5e-5bdb-8888-b6e3f3618c0a", 00:10:46.546 "is_configured": true, 00:10:46.546 "data_offset": 2048, 00:10:46.546 "data_size": 63488 00:10:46.546 }, 00:10:46.546 { 00:10:46.546 "name": "BaseBdev4", 00:10:46.546 "uuid": "e3220a28-c434-59ee-a241-e501660bd233", 00:10:46.546 "is_configured": true, 00:10:46.546 "data_offset": 2048, 00:10:46.546 "data_size": 63488 00:10:46.546 } 00:10:46.546 ] 00:10:46.546 }' 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.546 16:49:08 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:47.114 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:47.114 16:49:08 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:47.114 [2024-09-29 16:49:08.564967] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:48.050 "name": "raid_bdev1", 00:10:48.050 "uuid": "41bf150a-d3e1-42af-8420-0218d6725ca3", 00:10:48.050 "strip_size_kb": 0, 00:10:48.050 "state": "online", 00:10:48.050 "raid_level": "raid1", 00:10:48.050 "superblock": true, 00:10:48.050 "num_base_bdevs": 4, 00:10:48.050 "num_base_bdevs_discovered": 4, 00:10:48.050 "num_base_bdevs_operational": 4, 00:10:48.050 "base_bdevs_list": [ 00:10:48.050 { 00:10:48.050 "name": "BaseBdev1", 00:10:48.050 "uuid": "cbb3859b-06fd-5a76-8020-740ec659cefc", 00:10:48.050 "is_configured": true, 00:10:48.050 "data_offset": 2048, 00:10:48.050 "data_size": 63488 00:10:48.050 }, 00:10:48.050 { 00:10:48.050 "name": "BaseBdev2", 00:10:48.050 "uuid": "a6c9bf49-cd6d-533f-bda5-fe62fccb4681", 00:10:48.050 "is_configured": true, 00:10:48.050 "data_offset": 2048, 00:10:48.050 "data_size": 63488 00:10:48.050 }, 00:10:48.050 { 00:10:48.050 "name": "BaseBdev3", 00:10:48.050 "uuid": "c35c6a5a-ea5e-5bdb-8888-b6e3f3618c0a", 00:10:48.050 "is_configured": true, 00:10:48.050 "data_offset": 2048, 00:10:48.050 "data_size": 63488 00:10:48.050 }, 00:10:48.050 { 00:10:48.050 "name": "BaseBdev4", 00:10:48.050 "uuid": "e3220a28-c434-59ee-a241-e501660bd233", 00:10:48.050 "is_configured": true, 00:10:48.050 "data_offset": 2048, 00:10:48.050 "data_size": 63488 00:10:48.050 } 00:10:48.050 ] 00:10:48.050 }' 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:48.050 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.617 16:49:09 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:48.617 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:48.617 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.617 [2024-09-29 16:49:09.995765] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:48.617 [2024-09-29 16:49:09.995854] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:48.617 [2024-09-29 16:49:09.998392] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:48.617 [2024-09-29 16:49:09.998482] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:48.617 [2024-09-29 16:49:09.998650] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:48.617 [2024-09-29 16:49:09.998697] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:48.617 { 00:10:48.617 "results": [ 00:10:48.617 { 00:10:48.617 "job": "raid_bdev1", 00:10:48.617 "core_mask": "0x1", 00:10:48.617 "workload": "randrw", 00:10:48.617 "percentage": 50, 00:10:48.617 "status": "finished", 00:10:48.617 "queue_depth": 1, 00:10:48.617 "io_size": 131072, 00:10:48.617 "runtime": 1.431786, 00:10:48.617 "iops": 11710.548922813885, 00:10:48.617 "mibps": 1463.8186153517356, 00:10:48.617 "io_failed": 0, 00:10:48.617 "io_timeout": 0, 00:10:48.617 "avg_latency_us": 82.86908756881824, 00:10:48.617 "min_latency_us": 21.910917030567685, 00:10:48.617 "max_latency_us": 1523.926637554585 00:10:48.617 } 00:10:48.617 ], 00:10:48.617 "core_count": 1 00:10:48.617 } 00:10:48.617 16:49:09 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85423 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 85423 ']' 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 85423 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85423 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:48.617 killing process with pid 85423 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85423' 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 85423 00:10:48.617 [2024-09-29 16:49:10.044044] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:48.617 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 85423 00:10:48.617 [2024-09-29 16:49:10.078865] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.5J1Lz8YAhz 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:48.877 00:10:48.877 real 0m3.344s 00:10:48.877 user 0m4.184s 00:10:48.877 sys 0m0.551s 00:10:48.877 ************************************ 00:10:48.877 END TEST raid_read_error_test 00:10:48.877 ************************************ 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:48.877 16:49:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.877 16:49:10 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:10:48.877 16:49:10 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:48.877 16:49:10 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:48.877 16:49:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:48.877 ************************************ 00:10:48.877 START TEST raid_write_error_test 00:10:48.877 ************************************ 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 write 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.FgTjAQMESm 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85558 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85558 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 85558 ']' 00:10:48.877 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:48.877 16:49:10 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.877 [2024-09-29 16:49:10.495260] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:48.877 [2024-09-29 16:49:10.495404] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85558 ] 00:10:49.137 [2024-09-29 16:49:10.639922] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:49.137 [2024-09-29 16:49:10.685781] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:49.137 [2024-09-29 16:49:10.727501] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:49.137 [2024-09-29 16:49:10.727620] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.704 BaseBdev1_malloc 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.704 true 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.704 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:49.705 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.705 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.705 [2024-09-29 16:49:11.349500] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:49.705 [2024-09-29 16:49:11.349613] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.705 [2024-09-29 16:49:11.349657] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:49.705 [2024-09-29 16:49:11.349694] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.705 [2024-09-29 16:49:11.351850] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.705 [2024-09-29 16:49:11.351921] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:49.705 BaseBdev1 00:10:49.705 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.705 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:49.705 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:49.705 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.705 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.964 BaseBdev2_malloc 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.964 true 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.964 [2024-09-29 16:49:11.400591] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:49.964 [2024-09-29 16:49:11.400696] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.964 [2024-09-29 16:49:11.400719] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:49.964 [2024-09-29 16:49:11.400728] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.964 [2024-09-29 16:49:11.402789] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.964 [2024-09-29 16:49:11.402824] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:49.964 BaseBdev2 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.964 BaseBdev3_malloc 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.964 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.965 true 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.965 [2024-09-29 16:49:11.441072] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:49.965 [2024-09-29 16:49:11.441116] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.965 [2024-09-29 16:49:11.441133] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:49.965 [2024-09-29 16:49:11.441141] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.965 [2024-09-29 16:49:11.443176] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.965 [2024-09-29 16:49:11.443212] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:49.965 BaseBdev3 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.965 BaseBdev4_malloc 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.965 true 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.965 [2024-09-29 16:49:11.481532] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:49.965 [2024-09-29 16:49:11.481577] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.965 [2024-09-29 16:49:11.481597] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:49.965 [2024-09-29 16:49:11.481605] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.965 [2024-09-29 16:49:11.483625] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.965 [2024-09-29 16:49:11.483661] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:49.965 BaseBdev4 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.965 [2024-09-29 16:49:11.493566] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:49.965 [2024-09-29 16:49:11.495378] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:49.965 [2024-09-29 16:49:11.495516] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:49.965 [2024-09-29 16:49:11.495587] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:49.965 [2024-09-29 16:49:11.495796] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:49.965 [2024-09-29 16:49:11.495809] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:10:49.965 [2024-09-29 16:49:11.496053] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:49.965 [2024-09-29 16:49:11.496205] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:49.965 [2024-09-29 16:49:11.496218] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:49.965 [2024-09-29 16:49:11.496348] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.965 "name": "raid_bdev1", 00:10:49.965 "uuid": "545f05c9-b5b0-451a-8dac-1fadddb046cb", 00:10:49.965 "strip_size_kb": 0, 00:10:49.965 "state": "online", 00:10:49.965 "raid_level": "raid1", 00:10:49.965 "superblock": true, 00:10:49.965 "num_base_bdevs": 4, 00:10:49.965 "num_base_bdevs_discovered": 4, 00:10:49.965 "num_base_bdevs_operational": 4, 00:10:49.965 "base_bdevs_list": [ 00:10:49.965 { 00:10:49.965 "name": "BaseBdev1", 00:10:49.965 "uuid": "01425089-87a8-5c33-a246-d997770eaa04", 00:10:49.965 "is_configured": true, 00:10:49.965 "data_offset": 2048, 00:10:49.965 "data_size": 63488 00:10:49.965 }, 00:10:49.965 { 00:10:49.965 "name": "BaseBdev2", 00:10:49.965 "uuid": "4afe898f-14d5-576a-b64d-8a51f58fbfb6", 00:10:49.965 "is_configured": true, 00:10:49.965 "data_offset": 2048, 00:10:49.965 "data_size": 63488 00:10:49.965 }, 00:10:49.965 { 00:10:49.965 "name": "BaseBdev3", 00:10:49.965 "uuid": "2ff0b8b6-c312-5532-90a9-ffced41812e0", 00:10:49.965 "is_configured": true, 00:10:49.965 "data_offset": 2048, 00:10:49.965 "data_size": 63488 00:10:49.965 }, 00:10:49.965 { 00:10:49.965 "name": "BaseBdev4", 00:10:49.965 "uuid": "a3a2e785-01f0-5e38-bc48-9fd742bfc43f", 00:10:49.965 "is_configured": true, 00:10:49.965 "data_offset": 2048, 00:10:49.965 "data_size": 63488 00:10:49.965 } 00:10:49.965 ] 00:10:49.965 }' 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.965 16:49:11 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.534 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:50.534 16:49:11 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:50.534 [2024-09-29 16:49:11.989070] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.473 [2024-09-29 16:49:12.907186] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:10:51.473 [2024-09-29 16:49:12.907333] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:51.473 [2024-09-29 16:49:12.907593] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.473 "name": "raid_bdev1", 00:10:51.473 "uuid": "545f05c9-b5b0-451a-8dac-1fadddb046cb", 00:10:51.473 "strip_size_kb": 0, 00:10:51.473 "state": "online", 00:10:51.473 "raid_level": "raid1", 00:10:51.473 "superblock": true, 00:10:51.473 "num_base_bdevs": 4, 00:10:51.473 "num_base_bdevs_discovered": 3, 00:10:51.473 "num_base_bdevs_operational": 3, 00:10:51.473 "base_bdevs_list": [ 00:10:51.473 { 00:10:51.473 "name": null, 00:10:51.473 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:51.473 "is_configured": false, 00:10:51.473 "data_offset": 0, 00:10:51.473 "data_size": 63488 00:10:51.473 }, 00:10:51.473 { 00:10:51.473 "name": "BaseBdev2", 00:10:51.473 "uuid": "4afe898f-14d5-576a-b64d-8a51f58fbfb6", 00:10:51.473 "is_configured": true, 00:10:51.473 "data_offset": 2048, 00:10:51.473 "data_size": 63488 00:10:51.473 }, 00:10:51.473 { 00:10:51.473 "name": "BaseBdev3", 00:10:51.473 "uuid": "2ff0b8b6-c312-5532-90a9-ffced41812e0", 00:10:51.473 "is_configured": true, 00:10:51.473 "data_offset": 2048, 00:10:51.473 "data_size": 63488 00:10:51.473 }, 00:10:51.473 { 00:10:51.473 "name": "BaseBdev4", 00:10:51.473 "uuid": "a3a2e785-01f0-5e38-bc48-9fd742bfc43f", 00:10:51.473 "is_configured": true, 00:10:51.473 "data_offset": 2048, 00:10:51.473 "data_size": 63488 00:10:51.473 } 00:10:51.473 ] 00:10:51.473 }' 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.473 16:49:12 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.733 [2024-09-29 16:49:13.358517] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:51.733 [2024-09-29 16:49:13.358615] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:51.733 [2024-09-29 16:49:13.361054] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:51.733 [2024-09-29 16:49:13.361155] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:51.733 [2024-09-29 16:49:13.361314] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:51.733 [2024-09-29 16:49:13.361368] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:51.733 { 00:10:51.733 "results": [ 00:10:51.733 { 00:10:51.733 "job": "raid_bdev1", 00:10:51.733 "core_mask": "0x1", 00:10:51.733 "workload": "randrw", 00:10:51.733 "percentage": 50, 00:10:51.733 "status": "finished", 00:10:51.733 "queue_depth": 1, 00:10:51.733 "io_size": 131072, 00:10:51.733 "runtime": 1.370298, 00:10:51.733 "iops": 12737.375373823796, 00:10:51.733 "mibps": 1592.1719217279744, 00:10:51.733 "io_failed": 0, 00:10:51.733 "io_timeout": 0, 00:10:51.733 "avg_latency_us": 75.99041312835786, 00:10:51.733 "min_latency_us": 21.687336244541484, 00:10:51.733 "max_latency_us": 1395.1441048034935 00:10:51.733 } 00:10:51.733 ], 00:10:51.733 "core_count": 1 00:10:51.733 } 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85558 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 85558 ']' 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 85558 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85558 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85558' 00:10:51.733 killing process with pid 85558 00:10:51.733 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 85558 00:10:51.992 [2024-09-29 16:49:13.405561] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:51.992 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 85558 00:10:51.992 [2024-09-29 16:49:13.440158] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.FgTjAQMESm 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:52.251 ************************************ 00:10:52.251 END TEST raid_write_error_test 00:10:52.251 ************************************ 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:10:52.251 00:10:52.251 real 0m3.285s 00:10:52.251 user 0m4.109s 00:10:52.251 sys 0m0.526s 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:52.251 16:49:13 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.251 16:49:13 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:10:52.251 16:49:13 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:10:52.251 16:49:13 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:10:52.251 16:49:13 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:10:52.251 16:49:13 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:52.251 16:49:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:52.251 ************************************ 00:10:52.251 START TEST raid_rebuild_test 00:10:52.251 ************************************ 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false false true 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=85685 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 85685 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 85685 ']' 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:52.251 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:52.251 16:49:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.251 I/O size of 3145728 is greater than zero copy threshold (65536). 00:10:52.251 Zero copy mechanism will not be used. 00:10:52.251 [2024-09-29 16:49:13.851179] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:10:52.251 [2024-09-29 16:49:13.851294] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85685 ] 00:10:52.511 [2024-09-29 16:49:13.996059] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:52.511 [2024-09-29 16:49:14.041506] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:52.511 [2024-09-29 16:49:14.083266] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:52.511 [2024-09-29 16:49:14.083387] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.080 BaseBdev1_malloc 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.080 [2024-09-29 16:49:14.685405] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:10:53.080 [2024-09-29 16:49:14.685527] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.080 [2024-09-29 16:49:14.685570] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:53.080 [2024-09-29 16:49:14.685603] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.080 [2024-09-29 16:49:14.687758] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.080 [2024-09-29 16:49:14.687824] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:53.080 BaseBdev1 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.080 BaseBdev2_malloc 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.080 [2024-09-29 16:49:14.727595] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:10:53.080 [2024-09-29 16:49:14.727876] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.080 [2024-09-29 16:49:14.728043] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:53.080 [2024-09-29 16:49:14.728161] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.080 [2024-09-29 16:49:14.732635] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.080 [2024-09-29 16:49:14.732782] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:53.080 BaseBdev2 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.080 spare_malloc 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.080 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.340 spare_delay 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.340 [2024-09-29 16:49:14.770452] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:10:53.340 [2024-09-29 16:49:14.770503] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.340 [2024-09-29 16:49:14.770524] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:10:53.340 [2024-09-29 16:49:14.770531] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.340 [2024-09-29 16:49:14.772618] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.340 [2024-09-29 16:49:14.772654] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:10:53.340 spare 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.340 [2024-09-29 16:49:14.782481] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:53.340 [2024-09-29 16:49:14.784282] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:53.340 [2024-09-29 16:49:14.784374] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:53.340 [2024-09-29 16:49:14.784387] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:10:53.340 [2024-09-29 16:49:14.784650] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:10:53.340 [2024-09-29 16:49:14.784799] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:53.340 [2024-09-29 16:49:14.784812] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:53.340 [2024-09-29 16:49:14.784950] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.340 "name": "raid_bdev1", 00:10:53.340 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:10:53.340 "strip_size_kb": 0, 00:10:53.340 "state": "online", 00:10:53.340 "raid_level": "raid1", 00:10:53.340 "superblock": false, 00:10:53.340 "num_base_bdevs": 2, 00:10:53.340 "num_base_bdevs_discovered": 2, 00:10:53.340 "num_base_bdevs_operational": 2, 00:10:53.340 "base_bdevs_list": [ 00:10:53.340 { 00:10:53.340 "name": "BaseBdev1", 00:10:53.340 "uuid": "f846b183-2591-5963-883e-82b7b6c16df6", 00:10:53.340 "is_configured": true, 00:10:53.340 "data_offset": 0, 00:10:53.340 "data_size": 65536 00:10:53.340 }, 00:10:53.340 { 00:10:53.340 "name": "BaseBdev2", 00:10:53.340 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:10:53.340 "is_configured": true, 00:10:53.340 "data_offset": 0, 00:10:53.340 "data_size": 65536 00:10:53.340 } 00:10:53.340 ] 00:10:53.340 }' 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.340 16:49:14 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.600 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:10:53.600 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:53.600 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.600 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.600 [2024-09-29 16:49:15.257881] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:53.859 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:10:53.859 [2024-09-29 16:49:15.501290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:53.859 /dev/nbd0 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:10:54.119 1+0 records in 00:10:54.119 1+0 records out 00:10:54.119 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000525014 s, 7.8 MB/s 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:10:54.119 16:49:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:10:58.371 65536+0 records in 00:10:58.371 65536+0 records out 00:10:58.371 33554432 bytes (34 MB, 32 MiB) copied, 3.64873 s, 9.2 MB/s 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:10:58.371 [2024-09-29 16:49:19.441432] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.371 [2024-09-29 16:49:19.462040] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.371 "name": "raid_bdev1", 00:10:58.371 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:10:58.371 "strip_size_kb": 0, 00:10:58.371 "state": "online", 00:10:58.371 "raid_level": "raid1", 00:10:58.371 "superblock": false, 00:10:58.371 "num_base_bdevs": 2, 00:10:58.371 "num_base_bdevs_discovered": 1, 00:10:58.371 "num_base_bdevs_operational": 1, 00:10:58.371 "base_bdevs_list": [ 00:10:58.371 { 00:10:58.371 "name": null, 00:10:58.371 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:58.371 "is_configured": false, 00:10:58.371 "data_offset": 0, 00:10:58.371 "data_size": 65536 00:10:58.371 }, 00:10:58.371 { 00:10:58.371 "name": "BaseBdev2", 00:10:58.371 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:10:58.371 "is_configured": true, 00:10:58.371 "data_offset": 0, 00:10:58.371 "data_size": 65536 00:10:58.371 } 00:10:58.371 ] 00:10:58.371 }' 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.371 [2024-09-29 16:49:19.913275] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:10:58.371 [2024-09-29 16:49:19.917466] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06220 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:58.371 16:49:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:10:58.371 [2024-09-29 16:49:19.919423] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:59.309 16:49:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:59.310 16:49:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.310 16:49:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:59.310 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:10:59.310 "name": "raid_bdev1", 00:10:59.310 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:10:59.310 "strip_size_kb": 0, 00:10:59.310 "state": "online", 00:10:59.310 "raid_level": "raid1", 00:10:59.310 "superblock": false, 00:10:59.310 "num_base_bdevs": 2, 00:10:59.310 "num_base_bdevs_discovered": 2, 00:10:59.310 "num_base_bdevs_operational": 2, 00:10:59.310 "process": { 00:10:59.310 "type": "rebuild", 00:10:59.310 "target": "spare", 00:10:59.310 "progress": { 00:10:59.310 "blocks": 20480, 00:10:59.310 "percent": 31 00:10:59.310 } 00:10:59.310 }, 00:10:59.310 "base_bdevs_list": [ 00:10:59.310 { 00:10:59.310 "name": "spare", 00:10:59.310 "uuid": "3be5c419-55ff-59fc-901c-5f82be9d841f", 00:10:59.310 "is_configured": true, 00:10:59.310 "data_offset": 0, 00:10:59.310 "data_size": 65536 00:10:59.310 }, 00:10:59.310 { 00:10:59.310 "name": "BaseBdev2", 00:10:59.310 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:10:59.310 "is_configured": true, 00:10:59.310 "data_offset": 0, 00:10:59.310 "data_size": 65536 00:10:59.310 } 00:10:59.310 ] 00:10:59.310 }' 00:10:59.310 16:49:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.570 [2024-09-29 16:49:21.060127] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:59.570 [2024-09-29 16:49:21.124026] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:10:59.570 [2024-09-29 16:49:21.124144] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:59.570 [2024-09-29 16:49:21.124181] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:10:59.570 [2024-09-29 16:49:21.124202] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:59.570 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:59.571 "name": "raid_bdev1", 00:10:59.571 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:10:59.571 "strip_size_kb": 0, 00:10:59.571 "state": "online", 00:10:59.571 "raid_level": "raid1", 00:10:59.571 "superblock": false, 00:10:59.571 "num_base_bdevs": 2, 00:10:59.571 "num_base_bdevs_discovered": 1, 00:10:59.571 "num_base_bdevs_operational": 1, 00:10:59.571 "base_bdevs_list": [ 00:10:59.571 { 00:10:59.571 "name": null, 00:10:59.571 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:59.571 "is_configured": false, 00:10:59.571 "data_offset": 0, 00:10:59.571 "data_size": 65536 00:10:59.571 }, 00:10:59.571 { 00:10:59.571 "name": "BaseBdev2", 00:10:59.571 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:10:59.571 "is_configured": true, 00:10:59.571 "data_offset": 0, 00:10:59.571 "data_size": 65536 00:10:59.571 } 00:10:59.571 ] 00:10:59.571 }' 00:10:59.571 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:59.571 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:00.140 "name": "raid_bdev1", 00:11:00.140 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:11:00.140 "strip_size_kb": 0, 00:11:00.140 "state": "online", 00:11:00.140 "raid_level": "raid1", 00:11:00.140 "superblock": false, 00:11:00.140 "num_base_bdevs": 2, 00:11:00.140 "num_base_bdevs_discovered": 1, 00:11:00.140 "num_base_bdevs_operational": 1, 00:11:00.140 "base_bdevs_list": [ 00:11:00.140 { 00:11:00.140 "name": null, 00:11:00.140 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.140 "is_configured": false, 00:11:00.140 "data_offset": 0, 00:11:00.140 "data_size": 65536 00:11:00.140 }, 00:11:00.140 { 00:11:00.140 "name": "BaseBdev2", 00:11:00.140 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:11:00.140 "is_configured": true, 00:11:00.140 "data_offset": 0, 00:11:00.140 "data_size": 65536 00:11:00.140 } 00:11:00.140 ] 00:11:00.140 }' 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.140 [2024-09-29 16:49:21.759635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:00.140 [2024-09-29 16:49:21.763665] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d062f0 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.140 16:49:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:00.140 [2024-09-29 16:49:21.765555] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:01.523 "name": "raid_bdev1", 00:11:01.523 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:11:01.523 "strip_size_kb": 0, 00:11:01.523 "state": "online", 00:11:01.523 "raid_level": "raid1", 00:11:01.523 "superblock": false, 00:11:01.523 "num_base_bdevs": 2, 00:11:01.523 "num_base_bdevs_discovered": 2, 00:11:01.523 "num_base_bdevs_operational": 2, 00:11:01.523 "process": { 00:11:01.523 "type": "rebuild", 00:11:01.523 "target": "spare", 00:11:01.523 "progress": { 00:11:01.523 "blocks": 20480, 00:11:01.523 "percent": 31 00:11:01.523 } 00:11:01.523 }, 00:11:01.523 "base_bdevs_list": [ 00:11:01.523 { 00:11:01.523 "name": "spare", 00:11:01.523 "uuid": "3be5c419-55ff-59fc-901c-5f82be9d841f", 00:11:01.523 "is_configured": true, 00:11:01.523 "data_offset": 0, 00:11:01.523 "data_size": 65536 00:11:01.523 }, 00:11:01.523 { 00:11:01.523 "name": "BaseBdev2", 00:11:01.523 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:11:01.523 "is_configured": true, 00:11:01.523 "data_offset": 0, 00:11:01.523 "data_size": 65536 00:11:01.523 } 00:11:01.523 ] 00:11:01.523 }' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=287 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:01.523 "name": "raid_bdev1", 00:11:01.523 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:11:01.523 "strip_size_kb": 0, 00:11:01.523 "state": "online", 00:11:01.523 "raid_level": "raid1", 00:11:01.523 "superblock": false, 00:11:01.523 "num_base_bdevs": 2, 00:11:01.523 "num_base_bdevs_discovered": 2, 00:11:01.523 "num_base_bdevs_operational": 2, 00:11:01.523 "process": { 00:11:01.523 "type": "rebuild", 00:11:01.523 "target": "spare", 00:11:01.523 "progress": { 00:11:01.523 "blocks": 22528, 00:11:01.523 "percent": 34 00:11:01.523 } 00:11:01.523 }, 00:11:01.523 "base_bdevs_list": [ 00:11:01.523 { 00:11:01.523 "name": "spare", 00:11:01.523 "uuid": "3be5c419-55ff-59fc-901c-5f82be9d841f", 00:11:01.523 "is_configured": true, 00:11:01.523 "data_offset": 0, 00:11:01.523 "data_size": 65536 00:11:01.523 }, 00:11:01.523 { 00:11:01.523 "name": "BaseBdev2", 00:11:01.523 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:11:01.523 "is_configured": true, 00:11:01.523 "data_offset": 0, 00:11:01.523 "data_size": 65536 00:11:01.523 } 00:11:01.523 ] 00:11:01.523 }' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:01.523 16:49:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:01.523 16:49:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:01.523 16:49:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:02.464 "name": "raid_bdev1", 00:11:02.464 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:11:02.464 "strip_size_kb": 0, 00:11:02.464 "state": "online", 00:11:02.464 "raid_level": "raid1", 00:11:02.464 "superblock": false, 00:11:02.464 "num_base_bdevs": 2, 00:11:02.464 "num_base_bdevs_discovered": 2, 00:11:02.464 "num_base_bdevs_operational": 2, 00:11:02.464 "process": { 00:11:02.464 "type": "rebuild", 00:11:02.464 "target": "spare", 00:11:02.464 "progress": { 00:11:02.464 "blocks": 45056, 00:11:02.464 "percent": 68 00:11:02.464 } 00:11:02.464 }, 00:11:02.464 "base_bdevs_list": [ 00:11:02.464 { 00:11:02.464 "name": "spare", 00:11:02.464 "uuid": "3be5c419-55ff-59fc-901c-5f82be9d841f", 00:11:02.464 "is_configured": true, 00:11:02.464 "data_offset": 0, 00:11:02.464 "data_size": 65536 00:11:02.464 }, 00:11:02.464 { 00:11:02.464 "name": "BaseBdev2", 00:11:02.464 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:11:02.464 "is_configured": true, 00:11:02.464 "data_offset": 0, 00:11:02.464 "data_size": 65536 00:11:02.464 } 00:11:02.464 ] 00:11:02.464 }' 00:11:02.464 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:02.724 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:02.724 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:02.724 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:02.724 16:49:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:03.661 [2024-09-29 16:49:24.976463] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:03.661 [2024-09-29 16:49:24.976593] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:03.661 [2024-09-29 16:49:24.976639] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:03.661 "name": "raid_bdev1", 00:11:03.661 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:11:03.661 "strip_size_kb": 0, 00:11:03.661 "state": "online", 00:11:03.661 "raid_level": "raid1", 00:11:03.661 "superblock": false, 00:11:03.661 "num_base_bdevs": 2, 00:11:03.661 "num_base_bdevs_discovered": 2, 00:11:03.661 "num_base_bdevs_operational": 2, 00:11:03.661 "base_bdevs_list": [ 00:11:03.661 { 00:11:03.661 "name": "spare", 00:11:03.661 "uuid": "3be5c419-55ff-59fc-901c-5f82be9d841f", 00:11:03.661 "is_configured": true, 00:11:03.661 "data_offset": 0, 00:11:03.661 "data_size": 65536 00:11:03.661 }, 00:11:03.661 { 00:11:03.661 "name": "BaseBdev2", 00:11:03.661 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:11:03.661 "is_configured": true, 00:11:03.661 "data_offset": 0, 00:11:03.661 "data_size": 65536 00:11:03.661 } 00:11:03.661 ] 00:11:03.661 }' 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:03.661 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:03.922 "name": "raid_bdev1", 00:11:03.922 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:11:03.922 "strip_size_kb": 0, 00:11:03.922 "state": "online", 00:11:03.922 "raid_level": "raid1", 00:11:03.922 "superblock": false, 00:11:03.922 "num_base_bdevs": 2, 00:11:03.922 "num_base_bdevs_discovered": 2, 00:11:03.922 "num_base_bdevs_operational": 2, 00:11:03.922 "base_bdevs_list": [ 00:11:03.922 { 00:11:03.922 "name": "spare", 00:11:03.922 "uuid": "3be5c419-55ff-59fc-901c-5f82be9d841f", 00:11:03.922 "is_configured": true, 00:11:03.922 "data_offset": 0, 00:11:03.922 "data_size": 65536 00:11:03.922 }, 00:11:03.922 { 00:11:03.922 "name": "BaseBdev2", 00:11:03.922 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:11:03.922 "is_configured": true, 00:11:03.922 "data_offset": 0, 00:11:03.922 "data_size": 65536 00:11:03.922 } 00:11:03.922 ] 00:11:03.922 }' 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.922 "name": "raid_bdev1", 00:11:03.922 "uuid": "b9f7a309-694f-4c98-9256-6d1c2eb74090", 00:11:03.922 "strip_size_kb": 0, 00:11:03.922 "state": "online", 00:11:03.922 "raid_level": "raid1", 00:11:03.922 "superblock": false, 00:11:03.922 "num_base_bdevs": 2, 00:11:03.922 "num_base_bdevs_discovered": 2, 00:11:03.922 "num_base_bdevs_operational": 2, 00:11:03.922 "base_bdevs_list": [ 00:11:03.922 { 00:11:03.922 "name": "spare", 00:11:03.922 "uuid": "3be5c419-55ff-59fc-901c-5f82be9d841f", 00:11:03.922 "is_configured": true, 00:11:03.922 "data_offset": 0, 00:11:03.922 "data_size": 65536 00:11:03.922 }, 00:11:03.922 { 00:11:03.922 "name": "BaseBdev2", 00:11:03.922 "uuid": "ee95a18e-f0d7-5d37-805a-a5ece07e0bde", 00:11:03.922 "is_configured": true, 00:11:03.922 "data_offset": 0, 00:11:03.922 "data_size": 65536 00:11:03.922 } 00:11:03.922 ] 00:11:03.922 }' 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.922 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.493 [2024-09-29 16:49:25.927036] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:04.493 [2024-09-29 16:49:25.927062] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:04.493 [2024-09-29 16:49:25.927156] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:04.493 [2024-09-29 16:49:25.927218] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:04.493 [2024-09-29 16:49:25.927239] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:04.493 16:49:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:04.494 16:49:25 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:04.754 /dev/nbd0 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:04.754 1+0 records in 00:11:04.754 1+0 records out 00:11:04.754 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000313772 s, 13.1 MB/s 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:04.754 /dev/nbd1 00:11:04.754 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:05.015 1+0 records in 00:11:05.015 1+0 records out 00:11:05.015 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000420724 s, 9.7 MB/s 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:05.015 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 85685 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 85685 ']' 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 85685 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85685 00:11:05.275 killing process with pid 85685 00:11:05.275 Received shutdown signal, test time was about 60.000000 seconds 00:11:05.275 00:11:05.275 Latency(us) 00:11:05.275 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:05.275 =================================================================================================================== 00:11:05.275 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:05.275 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:05.276 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85685' 00:11:05.276 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 85685 00:11:05.276 [2024-09-29 16:49:26.936583] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:05.276 16:49:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 85685 00:11:05.536 [2024-09-29 16:49:26.967528] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:05.536 16:49:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:05.536 00:11:05.536 real 0m13.441s 00:11:05.536 user 0m15.601s 00:11:05.536 sys 0m2.641s 00:11:05.536 ************************************ 00:11:05.536 END TEST raid_rebuild_test 00:11:05.536 ************************************ 00:11:05.536 16:49:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:05.536 16:49:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.795 16:49:27 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:11:05.795 16:49:27 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:05.795 16:49:27 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:05.795 16:49:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:05.795 ************************************ 00:11:05.795 START TEST raid_rebuild_test_sb 00:11:05.795 ************************************ 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:05.795 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86084 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86084 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 86084 ']' 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:05.796 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:05.796 16:49:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:05.796 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:05.796 Zero copy mechanism will not be used. 00:11:05.796 [2024-09-29 16:49:27.363463] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:11:05.796 [2024-09-29 16:49:27.363597] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86084 ] 00:11:06.056 [2024-09-29 16:49:27.508624] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:06.056 [2024-09-29 16:49:27.553839] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:06.056 [2024-09-29 16:49:27.595720] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:06.056 [2024-09-29 16:49:27.595761] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.626 BaseBdev1_malloc 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:06.626 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.627 [2024-09-29 16:49:28.201740] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:06.627 [2024-09-29 16:49:28.201797] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.627 [2024-09-29 16:49:28.201835] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:06.627 [2024-09-29 16:49:28.201849] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.627 [2024-09-29 16:49:28.203975] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.627 [2024-09-29 16:49:28.204014] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:06.627 BaseBdev1 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.627 BaseBdev2_malloc 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.627 [2024-09-29 16:49:28.245347] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:06.627 [2024-09-29 16:49:28.245444] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.627 [2024-09-29 16:49:28.245490] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:06.627 [2024-09-29 16:49:28.245511] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.627 [2024-09-29 16:49:28.249673] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.627 [2024-09-29 16:49:28.249738] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:06.627 BaseBdev2 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.627 spare_malloc 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.627 spare_delay 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.627 [2024-09-29 16:49:28.287043] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:06.627 [2024-09-29 16:49:28.287094] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:06.627 [2024-09-29 16:49:28.287131] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:06.627 [2024-09-29 16:49:28.287139] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:06.627 [2024-09-29 16:49:28.289129] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:06.627 [2024-09-29 16:49:28.289163] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:06.627 spare 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.627 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.887 [2024-09-29 16:49:28.299073] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:06.887 [2024-09-29 16:49:28.300815] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:06.887 [2024-09-29 16:49:28.300990] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:06.887 [2024-09-29 16:49:28.301003] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:06.887 [2024-09-29 16:49:28.301308] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:06.887 [2024-09-29 16:49:28.301461] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:06.887 [2024-09-29 16:49:28.301479] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:06.887 [2024-09-29 16:49:28.301597] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.887 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.887 "name": "raid_bdev1", 00:11:06.887 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:06.887 "strip_size_kb": 0, 00:11:06.887 "state": "online", 00:11:06.887 "raid_level": "raid1", 00:11:06.887 "superblock": true, 00:11:06.887 "num_base_bdevs": 2, 00:11:06.887 "num_base_bdevs_discovered": 2, 00:11:06.887 "num_base_bdevs_operational": 2, 00:11:06.887 "base_bdevs_list": [ 00:11:06.887 { 00:11:06.887 "name": "BaseBdev1", 00:11:06.887 "uuid": "fbb83999-58be-5c9f-ae9e-0d8033cb6f43", 00:11:06.887 "is_configured": true, 00:11:06.887 "data_offset": 2048, 00:11:06.887 "data_size": 63488 00:11:06.887 }, 00:11:06.887 { 00:11:06.887 "name": "BaseBdev2", 00:11:06.887 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:06.887 "is_configured": true, 00:11:06.887 "data_offset": 2048, 00:11:06.888 "data_size": 63488 00:11:06.888 } 00:11:06.888 ] 00:11:06.888 }' 00:11:06.888 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.888 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:07.147 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:07.147 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.147 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:07.147 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:07.147 [2024-09-29 16:49:28.794442] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:07.147 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:07.407 16:49:28 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:07.407 [2024-09-29 16:49:29.045952] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:07.407 /dev/nbd0 00:11:07.407 16:49:29 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:07.668 1+0 records in 00:11:07.668 1+0 records out 00:11:07.668 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000354457 s, 11.6 MB/s 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:07.668 16:49:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:10.963 63488+0 records in 00:11:10.963 63488+0 records out 00:11:10.963 32505856 bytes (33 MB, 31 MiB) copied, 3.37465 s, 9.6 MB/s 00:11:10.963 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:10.963 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:10.963 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:10.963 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:10.963 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:10.963 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:10.963 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:11.223 [2024-09-29 16:49:32.695910] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.223 [2024-09-29 16:49:32.712298] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:11.223 "name": "raid_bdev1", 00:11:11.223 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:11.223 "strip_size_kb": 0, 00:11:11.223 "state": "online", 00:11:11.223 "raid_level": "raid1", 00:11:11.223 "superblock": true, 00:11:11.223 "num_base_bdevs": 2, 00:11:11.223 "num_base_bdevs_discovered": 1, 00:11:11.223 "num_base_bdevs_operational": 1, 00:11:11.223 "base_bdevs_list": [ 00:11:11.223 { 00:11:11.223 "name": null, 00:11:11.223 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.223 "is_configured": false, 00:11:11.223 "data_offset": 0, 00:11:11.223 "data_size": 63488 00:11:11.223 }, 00:11:11.223 { 00:11:11.223 "name": "BaseBdev2", 00:11:11.223 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:11.223 "is_configured": true, 00:11:11.223 "data_offset": 2048, 00:11:11.223 "data_size": 63488 00:11:11.223 } 00:11:11.223 ] 00:11:11.223 }' 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:11.223 16:49:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.482 16:49:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:11.482 16:49:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.482 16:49:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.482 [2024-09-29 16:49:33.123597] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:11.482 [2024-09-29 16:49:33.127878] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e280 00:11:11.482 16:49:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.482 16:49:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:11.482 [2024-09-29 16:49:33.129711] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:12.872 "name": "raid_bdev1", 00:11:12.872 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:12.872 "strip_size_kb": 0, 00:11:12.872 "state": "online", 00:11:12.872 "raid_level": "raid1", 00:11:12.872 "superblock": true, 00:11:12.872 "num_base_bdevs": 2, 00:11:12.872 "num_base_bdevs_discovered": 2, 00:11:12.872 "num_base_bdevs_operational": 2, 00:11:12.872 "process": { 00:11:12.872 "type": "rebuild", 00:11:12.872 "target": "spare", 00:11:12.872 "progress": { 00:11:12.872 "blocks": 20480, 00:11:12.872 "percent": 32 00:11:12.872 } 00:11:12.872 }, 00:11:12.872 "base_bdevs_list": [ 00:11:12.872 { 00:11:12.872 "name": "spare", 00:11:12.872 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:12.872 "is_configured": true, 00:11:12.872 "data_offset": 2048, 00:11:12.872 "data_size": 63488 00:11:12.872 }, 00:11:12.872 { 00:11:12.872 "name": "BaseBdev2", 00:11:12.872 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:12.872 "is_configured": true, 00:11:12.872 "data_offset": 2048, 00:11:12.872 "data_size": 63488 00:11:12.872 } 00:11:12.872 ] 00:11:12.872 }' 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.872 [2024-09-29 16:49:34.286523] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:12.872 [2024-09-29 16:49:34.334311] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:12.872 [2024-09-29 16:49:34.334384] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:12.872 [2024-09-29 16:49:34.334404] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:12.872 [2024-09-29 16:49:34.334411] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.872 "name": "raid_bdev1", 00:11:12.872 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:12.872 "strip_size_kb": 0, 00:11:12.872 "state": "online", 00:11:12.872 "raid_level": "raid1", 00:11:12.872 "superblock": true, 00:11:12.872 "num_base_bdevs": 2, 00:11:12.872 "num_base_bdevs_discovered": 1, 00:11:12.872 "num_base_bdevs_operational": 1, 00:11:12.872 "base_bdevs_list": [ 00:11:12.872 { 00:11:12.872 "name": null, 00:11:12.872 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.872 "is_configured": false, 00:11:12.872 "data_offset": 0, 00:11:12.872 "data_size": 63488 00:11:12.872 }, 00:11:12.872 { 00:11:12.872 "name": "BaseBdev2", 00:11:12.872 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:12.872 "is_configured": true, 00:11:12.872 "data_offset": 2048, 00:11:12.872 "data_size": 63488 00:11:12.872 } 00:11:12.872 ] 00:11:12.872 }' 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.872 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.130 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:13.130 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:13.130 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:13.130 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:13.130 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:13.130 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:13.130 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.131 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.131 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.131 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.131 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:13.131 "name": "raid_bdev1", 00:11:13.131 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:13.131 "strip_size_kb": 0, 00:11:13.131 "state": "online", 00:11:13.131 "raid_level": "raid1", 00:11:13.131 "superblock": true, 00:11:13.131 "num_base_bdevs": 2, 00:11:13.131 "num_base_bdevs_discovered": 1, 00:11:13.131 "num_base_bdevs_operational": 1, 00:11:13.131 "base_bdevs_list": [ 00:11:13.131 { 00:11:13.131 "name": null, 00:11:13.131 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:13.131 "is_configured": false, 00:11:13.131 "data_offset": 0, 00:11:13.131 "data_size": 63488 00:11:13.131 }, 00:11:13.131 { 00:11:13.131 "name": "BaseBdev2", 00:11:13.131 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:13.131 "is_configured": true, 00:11:13.131 "data_offset": 2048, 00:11:13.131 "data_size": 63488 00:11:13.131 } 00:11:13.131 ] 00:11:13.131 }' 00:11:13.131 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.390 [2024-09-29 16:49:34.905872] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:13.390 [2024-09-29 16:49:34.909929] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e350 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.390 16:49:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:13.390 [2024-09-29 16:49:34.911941] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:14.329 "name": "raid_bdev1", 00:11:14.329 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:14.329 "strip_size_kb": 0, 00:11:14.329 "state": "online", 00:11:14.329 "raid_level": "raid1", 00:11:14.329 "superblock": true, 00:11:14.329 "num_base_bdevs": 2, 00:11:14.329 "num_base_bdevs_discovered": 2, 00:11:14.329 "num_base_bdevs_operational": 2, 00:11:14.329 "process": { 00:11:14.329 "type": "rebuild", 00:11:14.329 "target": "spare", 00:11:14.329 "progress": { 00:11:14.329 "blocks": 20480, 00:11:14.329 "percent": 32 00:11:14.329 } 00:11:14.329 }, 00:11:14.329 "base_bdevs_list": [ 00:11:14.329 { 00:11:14.329 "name": "spare", 00:11:14.329 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:14.329 "is_configured": true, 00:11:14.329 "data_offset": 2048, 00:11:14.329 "data_size": 63488 00:11:14.329 }, 00:11:14.329 { 00:11:14.329 "name": "BaseBdev2", 00:11:14.329 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:14.329 "is_configured": true, 00:11:14.329 "data_offset": 2048, 00:11:14.329 "data_size": 63488 00:11:14.329 } 00:11:14.329 ] 00:11:14.329 }' 00:11:14.329 16:49:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:14.588 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:14.588 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:14.588 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:14.588 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:14.588 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:14.588 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:14.588 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=301 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:14.589 "name": "raid_bdev1", 00:11:14.589 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:14.589 "strip_size_kb": 0, 00:11:14.589 "state": "online", 00:11:14.589 "raid_level": "raid1", 00:11:14.589 "superblock": true, 00:11:14.589 "num_base_bdevs": 2, 00:11:14.589 "num_base_bdevs_discovered": 2, 00:11:14.589 "num_base_bdevs_operational": 2, 00:11:14.589 "process": { 00:11:14.589 "type": "rebuild", 00:11:14.589 "target": "spare", 00:11:14.589 "progress": { 00:11:14.589 "blocks": 22528, 00:11:14.589 "percent": 35 00:11:14.589 } 00:11:14.589 }, 00:11:14.589 "base_bdevs_list": [ 00:11:14.589 { 00:11:14.589 "name": "spare", 00:11:14.589 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:14.589 "is_configured": true, 00:11:14.589 "data_offset": 2048, 00:11:14.589 "data_size": 63488 00:11:14.589 }, 00:11:14.589 { 00:11:14.589 "name": "BaseBdev2", 00:11:14.589 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:14.589 "is_configured": true, 00:11:14.589 "data_offset": 2048, 00:11:14.589 "data_size": 63488 00:11:14.589 } 00:11:14.589 ] 00:11:14.589 }' 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:14.589 16:49:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:15.968 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:15.968 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:15.968 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:15.968 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:15.968 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:15.968 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:15.968 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:15.969 "name": "raid_bdev1", 00:11:15.969 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:15.969 "strip_size_kb": 0, 00:11:15.969 "state": "online", 00:11:15.969 "raid_level": "raid1", 00:11:15.969 "superblock": true, 00:11:15.969 "num_base_bdevs": 2, 00:11:15.969 "num_base_bdevs_discovered": 2, 00:11:15.969 "num_base_bdevs_operational": 2, 00:11:15.969 "process": { 00:11:15.969 "type": "rebuild", 00:11:15.969 "target": "spare", 00:11:15.969 "progress": { 00:11:15.969 "blocks": 47104, 00:11:15.969 "percent": 74 00:11:15.969 } 00:11:15.969 }, 00:11:15.969 "base_bdevs_list": [ 00:11:15.969 { 00:11:15.969 "name": "spare", 00:11:15.969 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:15.969 "is_configured": true, 00:11:15.969 "data_offset": 2048, 00:11:15.969 "data_size": 63488 00:11:15.969 }, 00:11:15.969 { 00:11:15.969 "name": "BaseBdev2", 00:11:15.969 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:15.969 "is_configured": true, 00:11:15.969 "data_offset": 2048, 00:11:15.969 "data_size": 63488 00:11:15.969 } 00:11:15.969 ] 00:11:15.969 }' 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:15.969 16:49:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:16.537 [2024-09-29 16:49:38.022362] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:16.537 [2024-09-29 16:49:38.022477] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:16.537 [2024-09-29 16:49:38.022639] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:16.795 "name": "raid_bdev1", 00:11:16.795 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:16.795 "strip_size_kb": 0, 00:11:16.795 "state": "online", 00:11:16.795 "raid_level": "raid1", 00:11:16.795 "superblock": true, 00:11:16.795 "num_base_bdevs": 2, 00:11:16.795 "num_base_bdevs_discovered": 2, 00:11:16.795 "num_base_bdevs_operational": 2, 00:11:16.795 "base_bdevs_list": [ 00:11:16.795 { 00:11:16.795 "name": "spare", 00:11:16.795 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:16.795 "is_configured": true, 00:11:16.795 "data_offset": 2048, 00:11:16.795 "data_size": 63488 00:11:16.795 }, 00:11:16.795 { 00:11:16.795 "name": "BaseBdev2", 00:11:16.795 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:16.795 "is_configured": true, 00:11:16.795 "data_offset": 2048, 00:11:16.795 "data_size": 63488 00:11:16.795 } 00:11:16.795 ] 00:11:16.795 }' 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:16.795 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:17.054 "name": "raid_bdev1", 00:11:17.054 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:17.054 "strip_size_kb": 0, 00:11:17.054 "state": "online", 00:11:17.054 "raid_level": "raid1", 00:11:17.054 "superblock": true, 00:11:17.054 "num_base_bdevs": 2, 00:11:17.054 "num_base_bdevs_discovered": 2, 00:11:17.054 "num_base_bdevs_operational": 2, 00:11:17.054 "base_bdevs_list": [ 00:11:17.054 { 00:11:17.054 "name": "spare", 00:11:17.054 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:17.054 "is_configured": true, 00:11:17.054 "data_offset": 2048, 00:11:17.054 "data_size": 63488 00:11:17.054 }, 00:11:17.054 { 00:11:17.054 "name": "BaseBdev2", 00:11:17.054 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:17.054 "is_configured": true, 00:11:17.054 "data_offset": 2048, 00:11:17.054 "data_size": 63488 00:11:17.054 } 00:11:17.054 ] 00:11:17.054 }' 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.054 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.054 "name": "raid_bdev1", 00:11:17.054 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:17.054 "strip_size_kb": 0, 00:11:17.054 "state": "online", 00:11:17.054 "raid_level": "raid1", 00:11:17.054 "superblock": true, 00:11:17.054 "num_base_bdevs": 2, 00:11:17.054 "num_base_bdevs_discovered": 2, 00:11:17.054 "num_base_bdevs_operational": 2, 00:11:17.054 "base_bdevs_list": [ 00:11:17.054 { 00:11:17.054 "name": "spare", 00:11:17.054 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:17.054 "is_configured": true, 00:11:17.054 "data_offset": 2048, 00:11:17.054 "data_size": 63488 00:11:17.054 }, 00:11:17.054 { 00:11:17.054 "name": "BaseBdev2", 00:11:17.054 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:17.054 "is_configured": true, 00:11:17.054 "data_offset": 2048, 00:11:17.055 "data_size": 63488 00:11:17.055 } 00:11:17.055 ] 00:11:17.055 }' 00:11:17.055 16:49:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.055 16:49:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.623 [2024-09-29 16:49:39.025004] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:17.623 [2024-09-29 16:49:39.025075] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:17.623 [2024-09-29 16:49:39.025211] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:17.623 [2024-09-29 16:49:39.025331] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:17.623 [2024-09-29 16:49:39.025399] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:17.623 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:17.882 /dev/nbd0 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:17.882 1+0 records in 00:11:17.882 1+0 records out 00:11:17.882 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000506746 s, 8.1 MB/s 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:17.882 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:18.141 /dev/nbd1 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:18.141 1+0 records in 00:11:18.141 1+0 records out 00:11:18.141 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000357193 s, 11.5 MB/s 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:18.141 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:18.402 16:49:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.402 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.662 [2024-09-29 16:49:40.088591] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:18.662 [2024-09-29 16:49:40.088691] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:18.662 [2024-09-29 16:49:40.088765] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:18.662 [2024-09-29 16:49:40.088802] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:18.662 [2024-09-29 16:49:40.091076] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:18.662 [2024-09-29 16:49:40.091156] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:18.662 [2024-09-29 16:49:40.091246] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:18.662 [2024-09-29 16:49:40.091306] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:18.662 [2024-09-29 16:49:40.091422] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:18.662 spare 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.662 [2024-09-29 16:49:40.191320] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:11:18.662 [2024-09-29 16:49:40.191382] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:18.662 [2024-09-29 16:49:40.191702] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cae960 00:11:18.662 [2024-09-29 16:49:40.191914] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:11:18.662 [2024-09-29 16:49:40.191966] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:11:18.662 [2024-09-29 16:49:40.192181] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:18.662 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:18.663 "name": "raid_bdev1", 00:11:18.663 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:18.663 "strip_size_kb": 0, 00:11:18.663 "state": "online", 00:11:18.663 "raid_level": "raid1", 00:11:18.663 "superblock": true, 00:11:18.663 "num_base_bdevs": 2, 00:11:18.663 "num_base_bdevs_discovered": 2, 00:11:18.663 "num_base_bdevs_operational": 2, 00:11:18.663 "base_bdevs_list": [ 00:11:18.663 { 00:11:18.663 "name": "spare", 00:11:18.663 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:18.663 "is_configured": true, 00:11:18.663 "data_offset": 2048, 00:11:18.663 "data_size": 63488 00:11:18.663 }, 00:11:18.663 { 00:11:18.663 "name": "BaseBdev2", 00:11:18.663 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:18.663 "is_configured": true, 00:11:18.663 "data_offset": 2048, 00:11:18.663 "data_size": 63488 00:11:18.663 } 00:11:18.663 ] 00:11:18.663 }' 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:18.663 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:19.230 "name": "raid_bdev1", 00:11:19.230 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:19.230 "strip_size_kb": 0, 00:11:19.230 "state": "online", 00:11:19.230 "raid_level": "raid1", 00:11:19.230 "superblock": true, 00:11:19.230 "num_base_bdevs": 2, 00:11:19.230 "num_base_bdevs_discovered": 2, 00:11:19.230 "num_base_bdevs_operational": 2, 00:11:19.230 "base_bdevs_list": [ 00:11:19.230 { 00:11:19.230 "name": "spare", 00:11:19.230 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:19.230 "is_configured": true, 00:11:19.230 "data_offset": 2048, 00:11:19.230 "data_size": 63488 00:11:19.230 }, 00:11:19.230 { 00:11:19.230 "name": "BaseBdev2", 00:11:19.230 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:19.230 "is_configured": true, 00:11:19.230 "data_offset": 2048, 00:11:19.230 "data_size": 63488 00:11:19.230 } 00:11:19.230 ] 00:11:19.230 }' 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.230 [2024-09-29 16:49:40.819359] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:19.230 "name": "raid_bdev1", 00:11:19.230 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:19.230 "strip_size_kb": 0, 00:11:19.230 "state": "online", 00:11:19.230 "raid_level": "raid1", 00:11:19.230 "superblock": true, 00:11:19.230 "num_base_bdevs": 2, 00:11:19.230 "num_base_bdevs_discovered": 1, 00:11:19.230 "num_base_bdevs_operational": 1, 00:11:19.230 "base_bdevs_list": [ 00:11:19.230 { 00:11:19.230 "name": null, 00:11:19.230 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:19.230 "is_configured": false, 00:11:19.230 "data_offset": 0, 00:11:19.230 "data_size": 63488 00:11:19.230 }, 00:11:19.230 { 00:11:19.230 "name": "BaseBdev2", 00:11:19.230 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:19.230 "is_configured": true, 00:11:19.230 "data_offset": 2048, 00:11:19.230 "data_size": 63488 00:11:19.230 } 00:11:19.230 ] 00:11:19.230 }' 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:19.230 16:49:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.800 16:49:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:19.800 16:49:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:19.800 16:49:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.800 [2024-09-29 16:49:41.306545] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:19.800 [2024-09-29 16:49:41.306804] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:19.800 [2024-09-29 16:49:41.306890] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:19.800 [2024-09-29 16:49:41.306970] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:19.800 [2024-09-29 16:49:41.311012] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caea30 00:11:19.800 16:49:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:19.800 16:49:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:19.800 [2024-09-29 16:49:41.312975] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:20.739 "name": "raid_bdev1", 00:11:20.739 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:20.739 "strip_size_kb": 0, 00:11:20.739 "state": "online", 00:11:20.739 "raid_level": "raid1", 00:11:20.739 "superblock": true, 00:11:20.739 "num_base_bdevs": 2, 00:11:20.739 "num_base_bdevs_discovered": 2, 00:11:20.739 "num_base_bdevs_operational": 2, 00:11:20.739 "process": { 00:11:20.739 "type": "rebuild", 00:11:20.739 "target": "spare", 00:11:20.739 "progress": { 00:11:20.739 "blocks": 20480, 00:11:20.739 "percent": 32 00:11:20.739 } 00:11:20.739 }, 00:11:20.739 "base_bdevs_list": [ 00:11:20.739 { 00:11:20.739 "name": "spare", 00:11:20.739 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:20.739 "is_configured": true, 00:11:20.739 "data_offset": 2048, 00:11:20.739 "data_size": 63488 00:11:20.739 }, 00:11:20.739 { 00:11:20.739 "name": "BaseBdev2", 00:11:20.739 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:20.739 "is_configured": true, 00:11:20.739 "data_offset": 2048, 00:11:20.739 "data_size": 63488 00:11:20.739 } 00:11:20.739 ] 00:11:20.739 }' 00:11:20.739 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.998 [2024-09-29 16:49:42.461921] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:20.998 [2024-09-29 16:49:42.517064] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:20.998 [2024-09-29 16:49:42.517119] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:20.998 [2024-09-29 16:49:42.517137] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:20.998 [2024-09-29 16:49:42.517144] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.998 "name": "raid_bdev1", 00:11:20.998 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:20.998 "strip_size_kb": 0, 00:11:20.998 "state": "online", 00:11:20.998 "raid_level": "raid1", 00:11:20.998 "superblock": true, 00:11:20.998 "num_base_bdevs": 2, 00:11:20.998 "num_base_bdevs_discovered": 1, 00:11:20.998 "num_base_bdevs_operational": 1, 00:11:20.998 "base_bdevs_list": [ 00:11:20.998 { 00:11:20.998 "name": null, 00:11:20.998 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:20.998 "is_configured": false, 00:11:20.998 "data_offset": 0, 00:11:20.998 "data_size": 63488 00:11:20.998 }, 00:11:20.998 { 00:11:20.998 "name": "BaseBdev2", 00:11:20.998 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:20.998 "is_configured": true, 00:11:20.998 "data_offset": 2048, 00:11:20.998 "data_size": 63488 00:11:20.998 } 00:11:20.998 ] 00:11:20.998 }' 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.998 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.567 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:21.567 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.567 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:21.567 [2024-09-29 16:49:42.984675] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:21.567 [2024-09-29 16:49:42.984746] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:21.567 [2024-09-29 16:49:42.984772] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:11:21.567 [2024-09-29 16:49:42.984781] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:21.567 [2024-09-29 16:49:42.985241] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:21.567 [2024-09-29 16:49:42.985265] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:21.567 [2024-09-29 16:49:42.985353] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:21.567 [2024-09-29 16:49:42.985365] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:21.567 [2024-09-29 16:49:42.985379] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:21.567 [2024-09-29 16:49:42.985411] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:21.567 spare 00:11:21.567 [2024-09-29 16:49:42.989324] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:11:21.567 16:49:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.567 16:49:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:21.567 [2024-09-29 16:49:42.991230] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:22.508 16:49:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:22.508 16:49:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:22.508 16:49:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:22.508 16:49:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:22.508 16:49:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:22.508 16:49:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.508 16:49:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:22.508 "name": "raid_bdev1", 00:11:22.508 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:22.508 "strip_size_kb": 0, 00:11:22.508 "state": "online", 00:11:22.508 "raid_level": "raid1", 00:11:22.508 "superblock": true, 00:11:22.508 "num_base_bdevs": 2, 00:11:22.508 "num_base_bdevs_discovered": 2, 00:11:22.508 "num_base_bdevs_operational": 2, 00:11:22.508 "process": { 00:11:22.508 "type": "rebuild", 00:11:22.508 "target": "spare", 00:11:22.508 "progress": { 00:11:22.508 "blocks": 20480, 00:11:22.508 "percent": 32 00:11:22.508 } 00:11:22.508 }, 00:11:22.508 "base_bdevs_list": [ 00:11:22.508 { 00:11:22.508 "name": "spare", 00:11:22.508 "uuid": "00a3d062-d65d-5158-9d2e-4d1d8ba836dc", 00:11:22.508 "is_configured": true, 00:11:22.508 "data_offset": 2048, 00:11:22.508 "data_size": 63488 00:11:22.508 }, 00:11:22.508 { 00:11:22.508 "name": "BaseBdev2", 00:11:22.508 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:22.508 "is_configured": true, 00:11:22.508 "data_offset": 2048, 00:11:22.508 "data_size": 63488 00:11:22.508 } 00:11:22.508 ] 00:11:22.508 }' 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.508 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.508 [2024-09-29 16:49:44.155486] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:22.768 [2024-09-29 16:49:44.195335] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:22.768 [2024-09-29 16:49:44.195395] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:22.768 [2024-09-29 16:49:44.195409] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:22.768 [2024-09-29 16:49:44.195419] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.768 "name": "raid_bdev1", 00:11:22.768 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:22.768 "strip_size_kb": 0, 00:11:22.768 "state": "online", 00:11:22.768 "raid_level": "raid1", 00:11:22.768 "superblock": true, 00:11:22.768 "num_base_bdevs": 2, 00:11:22.768 "num_base_bdevs_discovered": 1, 00:11:22.768 "num_base_bdevs_operational": 1, 00:11:22.768 "base_bdevs_list": [ 00:11:22.768 { 00:11:22.768 "name": null, 00:11:22.768 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:22.768 "is_configured": false, 00:11:22.768 "data_offset": 0, 00:11:22.768 "data_size": 63488 00:11:22.768 }, 00:11:22.768 { 00:11:22.768 "name": "BaseBdev2", 00:11:22.768 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:22.768 "is_configured": true, 00:11:22.768 "data_offset": 2048, 00:11:22.768 "data_size": 63488 00:11:22.768 } 00:11:22.768 ] 00:11:22.768 }' 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.768 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:23.028 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:23.288 "name": "raid_bdev1", 00:11:23.288 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:23.288 "strip_size_kb": 0, 00:11:23.288 "state": "online", 00:11:23.288 "raid_level": "raid1", 00:11:23.288 "superblock": true, 00:11:23.288 "num_base_bdevs": 2, 00:11:23.288 "num_base_bdevs_discovered": 1, 00:11:23.288 "num_base_bdevs_operational": 1, 00:11:23.288 "base_bdevs_list": [ 00:11:23.288 { 00:11:23.288 "name": null, 00:11:23.288 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.288 "is_configured": false, 00:11:23.288 "data_offset": 0, 00:11:23.288 "data_size": 63488 00:11:23.288 }, 00:11:23.288 { 00:11:23.288 "name": "BaseBdev2", 00:11:23.288 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:23.288 "is_configured": true, 00:11:23.288 "data_offset": 2048, 00:11:23.288 "data_size": 63488 00:11:23.288 } 00:11:23.288 ] 00:11:23.288 }' 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:23.288 [2024-09-29 16:49:44.810838] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:23.288 [2024-09-29 16:49:44.810914] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:23.288 [2024-09-29 16:49:44.810935] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:11:23.288 [2024-09-29 16:49:44.810946] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:23.288 [2024-09-29 16:49:44.811362] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:23.288 [2024-09-29 16:49:44.811388] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:23.288 [2024-09-29 16:49:44.811460] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:23.288 [2024-09-29 16:49:44.811491] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:23.288 [2024-09-29 16:49:44.811498] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:23.288 [2024-09-29 16:49:44.811510] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:23.288 BaseBdev1 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.288 16:49:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.228 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.228 "name": "raid_bdev1", 00:11:24.228 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:24.228 "strip_size_kb": 0, 00:11:24.228 "state": "online", 00:11:24.228 "raid_level": "raid1", 00:11:24.228 "superblock": true, 00:11:24.228 "num_base_bdevs": 2, 00:11:24.228 "num_base_bdevs_discovered": 1, 00:11:24.228 "num_base_bdevs_operational": 1, 00:11:24.228 "base_bdevs_list": [ 00:11:24.228 { 00:11:24.228 "name": null, 00:11:24.228 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.228 "is_configured": false, 00:11:24.228 "data_offset": 0, 00:11:24.228 "data_size": 63488 00:11:24.228 }, 00:11:24.228 { 00:11:24.228 "name": "BaseBdev2", 00:11:24.228 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:24.228 "is_configured": true, 00:11:24.228 "data_offset": 2048, 00:11:24.228 "data_size": 63488 00:11:24.229 } 00:11:24.229 ] 00:11:24.229 }' 00:11:24.229 16:49:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.229 16:49:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:24.799 "name": "raid_bdev1", 00:11:24.799 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:24.799 "strip_size_kb": 0, 00:11:24.799 "state": "online", 00:11:24.799 "raid_level": "raid1", 00:11:24.799 "superblock": true, 00:11:24.799 "num_base_bdevs": 2, 00:11:24.799 "num_base_bdevs_discovered": 1, 00:11:24.799 "num_base_bdevs_operational": 1, 00:11:24.799 "base_bdevs_list": [ 00:11:24.799 { 00:11:24.799 "name": null, 00:11:24.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.799 "is_configured": false, 00:11:24.799 "data_offset": 0, 00:11:24.799 "data_size": 63488 00:11:24.799 }, 00:11:24.799 { 00:11:24.799 "name": "BaseBdev2", 00:11:24.799 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:24.799 "is_configured": true, 00:11:24.799 "data_offset": 2048, 00:11:24.799 "data_size": 63488 00:11:24.799 } 00:11:24.799 ] 00:11:24.799 }' 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:24.799 [2024-09-29 16:49:46.388131] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:24.799 [2024-09-29 16:49:46.388298] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:24.799 [2024-09-29 16:49:46.388311] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:24.799 request: 00:11:24.799 { 00:11:24.799 "base_bdev": "BaseBdev1", 00:11:24.799 "raid_bdev": "raid_bdev1", 00:11:24.799 "method": "bdev_raid_add_base_bdev", 00:11:24.799 "req_id": 1 00:11:24.799 } 00:11:24.799 Got JSON-RPC error response 00:11:24.799 response: 00:11:24.799 { 00:11:24.799 "code": -22, 00:11:24.799 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:24.799 } 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:24.799 16:49:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.739 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:25.999 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.999 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.999 "name": "raid_bdev1", 00:11:25.999 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:25.999 "strip_size_kb": 0, 00:11:25.999 "state": "online", 00:11:25.999 "raid_level": "raid1", 00:11:25.999 "superblock": true, 00:11:25.999 "num_base_bdevs": 2, 00:11:25.999 "num_base_bdevs_discovered": 1, 00:11:25.999 "num_base_bdevs_operational": 1, 00:11:25.999 "base_bdevs_list": [ 00:11:25.999 { 00:11:25.999 "name": null, 00:11:25.999 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.999 "is_configured": false, 00:11:25.999 "data_offset": 0, 00:11:25.999 "data_size": 63488 00:11:25.999 }, 00:11:25.999 { 00:11:25.999 "name": "BaseBdev2", 00:11:25.999 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:25.999 "is_configured": true, 00:11:25.999 "data_offset": 2048, 00:11:25.999 "data_size": 63488 00:11:25.999 } 00:11:25.999 ] 00:11:25.999 }' 00:11:25.999 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.999 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:26.259 "name": "raid_bdev1", 00:11:26.259 "uuid": "8140a95a-902e-43f5-9b82-093f078f9e93", 00:11:26.259 "strip_size_kb": 0, 00:11:26.259 "state": "online", 00:11:26.259 "raid_level": "raid1", 00:11:26.259 "superblock": true, 00:11:26.259 "num_base_bdevs": 2, 00:11:26.259 "num_base_bdevs_discovered": 1, 00:11:26.259 "num_base_bdevs_operational": 1, 00:11:26.259 "base_bdevs_list": [ 00:11:26.259 { 00:11:26.259 "name": null, 00:11:26.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:26.259 "is_configured": false, 00:11:26.259 "data_offset": 0, 00:11:26.259 "data_size": 63488 00:11:26.259 }, 00:11:26.259 { 00:11:26.259 "name": "BaseBdev2", 00:11:26.259 "uuid": "87ce54db-67f2-558d-9372-2c545017587a", 00:11:26.259 "is_configured": true, 00:11:26.259 "data_offset": 2048, 00:11:26.259 "data_size": 63488 00:11:26.259 } 00:11:26.259 ] 00:11:26.259 }' 00:11:26.259 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86084 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 86084 ']' 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 86084 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:26.520 16:49:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 86084 00:11:26.520 killing process with pid 86084 00:11:26.520 Received shutdown signal, test time was about 60.000000 seconds 00:11:26.520 00:11:26.520 Latency(us) 00:11:26.520 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:26.520 =================================================================================================================== 00:11:26.520 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:26.520 16:49:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:26.520 16:49:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:26.520 16:49:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 86084' 00:11:26.520 16:49:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 86084 00:11:26.520 [2024-09-29 16:49:48.035367] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:26.520 [2024-09-29 16:49:48.035490] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:26.520 16:49:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 86084 00:11:26.520 [2024-09-29 16:49:48.035542] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:26.520 [2024-09-29 16:49:48.035551] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:11:26.520 [2024-09-29 16:49:48.066940] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:11:26.789 00:11:26.789 real 0m21.027s 00:11:26.789 user 0m26.470s 00:11:26.789 sys 0m3.250s 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:26.789 ************************************ 00:11:26.789 END TEST raid_rebuild_test_sb 00:11:26.789 ************************************ 00:11:26.789 16:49:48 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:11:26.789 16:49:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:26.789 16:49:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:26.789 16:49:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:26.789 ************************************ 00:11:26.789 START TEST raid_rebuild_test_io 00:11:26.789 ************************************ 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false true true 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:26.789 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=86792 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 86792 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 86792 ']' 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:26.790 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:26.790 16:49:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.068 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:27.068 Zero copy mechanism will not be used. 00:11:27.068 [2024-09-29 16:49:48.476633] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:11:27.068 [2024-09-29 16:49:48.476784] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86792 ] 00:11:27.068 [2024-09-29 16:49:48.622561] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:27.068 [2024-09-29 16:49:48.669288] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:27.068 [2024-09-29 16:49:48.712446] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:27.068 [2024-09-29 16:49:48.712565] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.675 BaseBdev1_malloc 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.675 [2024-09-29 16:49:49.307541] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:27.675 [2024-09-29 16:49:49.307647] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.675 [2024-09-29 16:49:49.307688] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:27.675 [2024-09-29 16:49:49.307749] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.675 [2024-09-29 16:49:49.310026] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.675 [2024-09-29 16:49:49.310099] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:27.675 BaseBdev1 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.675 BaseBdev2_malloc 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.675 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.935 [2024-09-29 16:49:49.349872] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:27.935 [2024-09-29 16:49:49.350058] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.935 [2024-09-29 16:49:49.350147] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:27.935 [2024-09-29 16:49:49.350241] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.935 [2024-09-29 16:49:49.355153] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.935 [2024-09-29 16:49:49.355300] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:27.935 BaseBdev2 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.935 spare_malloc 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.935 spare_delay 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.935 [2024-09-29 16:49:49.393515] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:27.935 [2024-09-29 16:49:49.393566] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.935 [2024-09-29 16:49:49.393603] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:27.935 [2024-09-29 16:49:49.393610] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.935 [2024-09-29 16:49:49.395729] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.935 [2024-09-29 16:49:49.395772] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:27.935 spare 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.935 [2024-09-29 16:49:49.405552] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:27.935 [2024-09-29 16:49:49.407436] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:27.935 [2024-09-29 16:49:49.407594] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:27.935 [2024-09-29 16:49:49.407612] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:27.935 [2024-09-29 16:49:49.407893] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:27.935 [2024-09-29 16:49:49.408035] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:27.935 [2024-09-29 16:49:49.408048] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:27.935 [2024-09-29 16:49:49.408194] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.935 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:27.936 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.936 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:27.936 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.936 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:27.936 "name": "raid_bdev1", 00:11:27.936 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:27.936 "strip_size_kb": 0, 00:11:27.936 "state": "online", 00:11:27.936 "raid_level": "raid1", 00:11:27.936 "superblock": false, 00:11:27.936 "num_base_bdevs": 2, 00:11:27.936 "num_base_bdevs_discovered": 2, 00:11:27.936 "num_base_bdevs_operational": 2, 00:11:27.936 "base_bdevs_list": [ 00:11:27.936 { 00:11:27.936 "name": "BaseBdev1", 00:11:27.936 "uuid": "5b10e86b-387d-5eb7-9638-8159f69454dd", 00:11:27.936 "is_configured": true, 00:11:27.936 "data_offset": 0, 00:11:27.936 "data_size": 65536 00:11:27.936 }, 00:11:27.936 { 00:11:27.936 "name": "BaseBdev2", 00:11:27.936 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:27.936 "is_configured": true, 00:11:27.936 "data_offset": 0, 00:11:27.936 "data_size": 65536 00:11:27.936 } 00:11:27.936 ] 00:11:27.936 }' 00:11:27.936 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:27.936 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:28.506 [2024-09-29 16:49:49.880950] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.506 [2024-09-29 16:49:49.984504] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.506 16:49:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.506 16:49:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:28.506 16:49:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:28.506 "name": "raid_bdev1", 00:11:28.506 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:28.506 "strip_size_kb": 0, 00:11:28.506 "state": "online", 00:11:28.506 "raid_level": "raid1", 00:11:28.506 "superblock": false, 00:11:28.506 "num_base_bdevs": 2, 00:11:28.506 "num_base_bdevs_discovered": 1, 00:11:28.506 "num_base_bdevs_operational": 1, 00:11:28.506 "base_bdevs_list": [ 00:11:28.506 { 00:11:28.506 "name": null, 00:11:28.506 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:28.506 "is_configured": false, 00:11:28.506 "data_offset": 0, 00:11:28.506 "data_size": 65536 00:11:28.506 }, 00:11:28.506 { 00:11:28.506 "name": "BaseBdev2", 00:11:28.506 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:28.506 "is_configured": true, 00:11:28.506 "data_offset": 0, 00:11:28.506 "data_size": 65536 00:11:28.506 } 00:11:28.506 ] 00:11:28.506 }' 00:11:28.506 16:49:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:28.506 16:49:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.506 [2024-09-29 16:49:50.074477] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:28.506 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:28.506 Zero copy mechanism will not be used. 00:11:28.506 Running I/O for 60 seconds... 00:11:28.766 16:49:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:28.766 16:49:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:28.766 16:49:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:28.766 [2024-09-29 16:49:50.417549] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:29.025 16:49:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.025 16:49:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:29.025 [2024-09-29 16:49:50.459030] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:29.025 [2024-09-29 16:49:50.460954] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:29.025 [2024-09-29 16:49:50.573875] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:29.025 [2024-09-29 16:49:50.574326] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:29.025 [2024-09-29 16:49:50.693236] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:29.025 [2024-09-29 16:49:50.693453] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:29.285 [2024-09-29 16:49:50.910483] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:29.545 210.00 IOPS, 630.00 MiB/s [2024-09-29 16:49:51.118362] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:29.805 [2024-09-29 16:49:51.362527] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:29.805 [2024-09-29 16:49:51.363048] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.805 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.064 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:30.065 "name": "raid_bdev1", 00:11:30.065 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:30.065 "strip_size_kb": 0, 00:11:30.065 "state": "online", 00:11:30.065 "raid_level": "raid1", 00:11:30.065 "superblock": false, 00:11:30.065 "num_base_bdevs": 2, 00:11:30.065 "num_base_bdevs_discovered": 2, 00:11:30.065 "num_base_bdevs_operational": 2, 00:11:30.065 "process": { 00:11:30.065 "type": "rebuild", 00:11:30.065 "target": "spare", 00:11:30.065 "progress": { 00:11:30.065 "blocks": 14336, 00:11:30.065 "percent": 21 00:11:30.065 } 00:11:30.065 }, 00:11:30.065 "base_bdevs_list": [ 00:11:30.065 { 00:11:30.065 "name": "spare", 00:11:30.065 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:30.065 "is_configured": true, 00:11:30.065 "data_offset": 0, 00:11:30.065 "data_size": 65536 00:11:30.065 }, 00:11:30.065 { 00:11:30.065 "name": "BaseBdev2", 00:11:30.065 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:30.065 "is_configured": true, 00:11:30.065 "data_offset": 0, 00:11:30.065 "data_size": 65536 00:11:30.065 } 00:11:30.065 ] 00:11:30.065 }' 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:30.065 [2024-09-29 16:49:51.577575] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.065 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.065 [2024-09-29 16:49:51.591849] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:30.325 [2024-09-29 16:49:51.792606] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:30.325 [2024-09-29 16:49:51.799913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:30.325 [2024-09-29 16:49:51.799944] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:30.325 [2024-09-29 16:49:51.799957] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:30.325 [2024-09-29 16:49:51.821106] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:30.325 "name": "raid_bdev1", 00:11:30.325 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:30.325 "strip_size_kb": 0, 00:11:30.325 "state": "online", 00:11:30.325 "raid_level": "raid1", 00:11:30.325 "superblock": false, 00:11:30.325 "num_base_bdevs": 2, 00:11:30.325 "num_base_bdevs_discovered": 1, 00:11:30.325 "num_base_bdevs_operational": 1, 00:11:30.325 "base_bdevs_list": [ 00:11:30.325 { 00:11:30.325 "name": null, 00:11:30.325 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:30.325 "is_configured": false, 00:11:30.325 "data_offset": 0, 00:11:30.325 "data_size": 65536 00:11:30.325 }, 00:11:30.325 { 00:11:30.325 "name": "BaseBdev2", 00:11:30.325 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:30.325 "is_configured": true, 00:11:30.325 "data_offset": 0, 00:11:30.325 "data_size": 65536 00:11:30.325 } 00:11:30.325 ] 00:11:30.325 }' 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:30.325 16:49:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.847 178.50 IOPS, 535.50 MiB/s 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:30.847 "name": "raid_bdev1", 00:11:30.847 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:30.847 "strip_size_kb": 0, 00:11:30.847 "state": "online", 00:11:30.847 "raid_level": "raid1", 00:11:30.847 "superblock": false, 00:11:30.847 "num_base_bdevs": 2, 00:11:30.847 "num_base_bdevs_discovered": 1, 00:11:30.847 "num_base_bdevs_operational": 1, 00:11:30.847 "base_bdevs_list": [ 00:11:30.847 { 00:11:30.847 "name": null, 00:11:30.847 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:30.847 "is_configured": false, 00:11:30.847 "data_offset": 0, 00:11:30.847 "data_size": 65536 00:11:30.847 }, 00:11:30.847 { 00:11:30.847 "name": "BaseBdev2", 00:11:30.847 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:30.847 "is_configured": true, 00:11:30.847 "data_offset": 0, 00:11:30.847 "data_size": 65536 00:11:30.847 } 00:11:30.847 ] 00:11:30.847 }' 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:30.847 [2024-09-29 16:49:52.387163] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:30.847 16:49:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:30.847 [2024-09-29 16:49:52.429344] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:11:30.847 [2024-09-29 16:49:52.431290] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:31.107 [2024-09-29 16:49:52.538703] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:31.107 [2024-09-29 16:49:52.539196] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:31.107 [2024-09-29 16:49:52.740894] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:31.107 [2024-09-29 16:49:52.741048] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:31.368 [2024-09-29 16:49:52.975523] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:31.628 187.00 IOPS, 561.00 MiB/s [2024-09-29 16:49:53.200407] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:31.628 [2024-09-29 16:49:53.200804] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:31.888 "name": "raid_bdev1", 00:11:31.888 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:31.888 "strip_size_kb": 0, 00:11:31.888 "state": "online", 00:11:31.888 "raid_level": "raid1", 00:11:31.888 "superblock": false, 00:11:31.888 "num_base_bdevs": 2, 00:11:31.888 "num_base_bdevs_discovered": 2, 00:11:31.888 "num_base_bdevs_operational": 2, 00:11:31.888 "process": { 00:11:31.888 "type": "rebuild", 00:11:31.888 "target": "spare", 00:11:31.888 "progress": { 00:11:31.888 "blocks": 12288, 00:11:31.888 "percent": 18 00:11:31.888 } 00:11:31.888 }, 00:11:31.888 "base_bdevs_list": [ 00:11:31.888 { 00:11:31.888 "name": "spare", 00:11:31.888 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:31.888 "is_configured": true, 00:11:31.888 "data_offset": 0, 00:11:31.888 "data_size": 65536 00:11:31.888 }, 00:11:31.888 { 00:11:31.888 "name": "BaseBdev2", 00:11:31.888 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:31.888 "is_configured": true, 00:11:31.888 "data_offset": 0, 00:11:31.888 "data_size": 65536 00:11:31.888 } 00:11:31.888 ] 00:11:31.888 }' 00:11:31.888 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:31.888 [2024-09-29 16:49:53.516349] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:31.889 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:31.889 [2024-09-29 16:49:53.516761] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:31.889 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=318 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:32.148 "name": "raid_bdev1", 00:11:32.148 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:32.148 "strip_size_kb": 0, 00:11:32.148 "state": "online", 00:11:32.148 "raid_level": "raid1", 00:11:32.148 "superblock": false, 00:11:32.148 "num_base_bdevs": 2, 00:11:32.148 "num_base_bdevs_discovered": 2, 00:11:32.148 "num_base_bdevs_operational": 2, 00:11:32.148 "process": { 00:11:32.148 "type": "rebuild", 00:11:32.148 "target": "spare", 00:11:32.148 "progress": { 00:11:32.148 "blocks": 14336, 00:11:32.148 "percent": 21 00:11:32.148 } 00:11:32.148 }, 00:11:32.148 "base_bdevs_list": [ 00:11:32.148 { 00:11:32.148 "name": "spare", 00:11:32.148 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:32.148 "is_configured": true, 00:11:32.148 "data_offset": 0, 00:11:32.148 "data_size": 65536 00:11:32.148 }, 00:11:32.148 { 00:11:32.148 "name": "BaseBdev2", 00:11:32.148 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:32.148 "is_configured": true, 00:11:32.148 "data_offset": 0, 00:11:32.148 "data_size": 65536 00:11:32.148 } 00:11:32.148 ] 00:11:32.148 }' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:32.148 16:49:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:32.148 [2024-09-29 16:49:53.737064] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:32.148 [2024-09-29 16:49:53.737275] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:33.348 152.75 IOPS, 458.25 MiB/s [2024-09-29 16:49:54.715844] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:33.348 "name": "raid_bdev1", 00:11:33.348 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:33.348 "strip_size_kb": 0, 00:11:33.348 "state": "online", 00:11:33.348 "raid_level": "raid1", 00:11:33.348 "superblock": false, 00:11:33.348 "num_base_bdevs": 2, 00:11:33.348 "num_base_bdevs_discovered": 2, 00:11:33.348 "num_base_bdevs_operational": 2, 00:11:33.348 "process": { 00:11:33.348 "type": "rebuild", 00:11:33.348 "target": "spare", 00:11:33.348 "progress": { 00:11:33.348 "blocks": 32768, 00:11:33.348 "percent": 50 00:11:33.348 } 00:11:33.348 }, 00:11:33.348 "base_bdevs_list": [ 00:11:33.348 { 00:11:33.348 "name": "spare", 00:11:33.348 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:33.348 "is_configured": true, 00:11:33.348 "data_offset": 0, 00:11:33.348 "data_size": 65536 00:11:33.348 }, 00:11:33.348 { 00:11:33.348 "name": "BaseBdev2", 00:11:33.348 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:33.348 "is_configured": true, 00:11:33.348 "data_offset": 0, 00:11:33.348 "data_size": 65536 00:11:33.348 } 00:11:33.348 ] 00:11:33.348 }' 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:33.348 [2024-09-29 16:49:54.829240] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:33.348 16:49:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:33.867 134.20 IOPS, 402.60 MiB/s [2024-09-29 16:49:55.400533] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:11:34.437 [2024-09-29 16:49:55.825443] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:11:34.437 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:34.437 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:34.437 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:34.437 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:34.437 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:34.437 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:34.438 "name": "raid_bdev1", 00:11:34.438 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:34.438 "strip_size_kb": 0, 00:11:34.438 "state": "online", 00:11:34.438 "raid_level": "raid1", 00:11:34.438 "superblock": false, 00:11:34.438 "num_base_bdevs": 2, 00:11:34.438 "num_base_bdevs_discovered": 2, 00:11:34.438 "num_base_bdevs_operational": 2, 00:11:34.438 "process": { 00:11:34.438 "type": "rebuild", 00:11:34.438 "target": "spare", 00:11:34.438 "progress": { 00:11:34.438 "blocks": 53248, 00:11:34.438 "percent": 81 00:11:34.438 } 00:11:34.438 }, 00:11:34.438 "base_bdevs_list": [ 00:11:34.438 { 00:11:34.438 "name": "spare", 00:11:34.438 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:34.438 "is_configured": true, 00:11:34.438 "data_offset": 0, 00:11:34.438 "data_size": 65536 00:11:34.438 }, 00:11:34.438 { 00:11:34.438 "name": "BaseBdev2", 00:11:34.438 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:34.438 "is_configured": true, 00:11:34.438 "data_offset": 0, 00:11:34.438 "data_size": 65536 00:11:34.438 } 00:11:34.438 ] 00:11:34.438 }' 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:34.438 16:49:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:34.438 16:49:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:34.438 16:49:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:34.698 120.33 IOPS, 361.00 MiB/s [2024-09-29 16:49:56.145461] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:11:34.698 [2024-09-29 16:49:56.249763] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:11:35.268 [2024-09-29 16:49:56.682774] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:35.268 [2024-09-29 16:49:56.782637] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:35.268 [2024-09-29 16:49:56.790096] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:35.527 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:35.527 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:35.527 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:35.527 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:35.527 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:35.527 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:35.527 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:35.528 "name": "raid_bdev1", 00:11:35.528 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:35.528 "strip_size_kb": 0, 00:11:35.528 "state": "online", 00:11:35.528 "raid_level": "raid1", 00:11:35.528 "superblock": false, 00:11:35.528 "num_base_bdevs": 2, 00:11:35.528 "num_base_bdevs_discovered": 2, 00:11:35.528 "num_base_bdevs_operational": 2, 00:11:35.528 "base_bdevs_list": [ 00:11:35.528 { 00:11:35.528 "name": "spare", 00:11:35.528 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:35.528 "is_configured": true, 00:11:35.528 "data_offset": 0, 00:11:35.528 "data_size": 65536 00:11:35.528 }, 00:11:35.528 { 00:11:35.528 "name": "BaseBdev2", 00:11:35.528 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:35.528 "is_configured": true, 00:11:35.528 "data_offset": 0, 00:11:35.528 "data_size": 65536 00:11:35.528 } 00:11:35.528 ] 00:11:35.528 }' 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:35.528 109.00 IOPS, 327.00 MiB/s 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:35.528 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:35.788 "name": "raid_bdev1", 00:11:35.788 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:35.788 "strip_size_kb": 0, 00:11:35.788 "state": "online", 00:11:35.788 "raid_level": "raid1", 00:11:35.788 "superblock": false, 00:11:35.788 "num_base_bdevs": 2, 00:11:35.788 "num_base_bdevs_discovered": 2, 00:11:35.788 "num_base_bdevs_operational": 2, 00:11:35.788 "base_bdevs_list": [ 00:11:35.788 { 00:11:35.788 "name": "spare", 00:11:35.788 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:35.788 "is_configured": true, 00:11:35.788 "data_offset": 0, 00:11:35.788 "data_size": 65536 00:11:35.788 }, 00:11:35.788 { 00:11:35.788 "name": "BaseBdev2", 00:11:35.788 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:35.788 "is_configured": true, 00:11:35.788 "data_offset": 0, 00:11:35.788 "data_size": 65536 00:11:35.788 } 00:11:35.788 ] 00:11:35.788 }' 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:35.788 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:35.789 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.789 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:35.789 "name": "raid_bdev1", 00:11:35.789 "uuid": "4b1dba68-f8c6-4e3e-8091-9e11e88f0470", 00:11:35.789 "strip_size_kb": 0, 00:11:35.789 "state": "online", 00:11:35.789 "raid_level": "raid1", 00:11:35.789 "superblock": false, 00:11:35.789 "num_base_bdevs": 2, 00:11:35.789 "num_base_bdevs_discovered": 2, 00:11:35.789 "num_base_bdevs_operational": 2, 00:11:35.789 "base_bdevs_list": [ 00:11:35.789 { 00:11:35.789 "name": "spare", 00:11:35.789 "uuid": "ea8ab43f-b940-5567-820a-3ae820683982", 00:11:35.789 "is_configured": true, 00:11:35.789 "data_offset": 0, 00:11:35.789 "data_size": 65536 00:11:35.789 }, 00:11:35.789 { 00:11:35.789 "name": "BaseBdev2", 00:11:35.789 "uuid": "5d9cfd87-3bf1-5bc0-aa47-21fc1c93a5f6", 00:11:35.789 "is_configured": true, 00:11:35.789 "data_offset": 0, 00:11:35.789 "data_size": 65536 00:11:35.789 } 00:11:35.789 ] 00:11:35.789 }' 00:11:35.789 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:35.789 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.359 [2024-09-29 16:49:57.793917] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:36.359 [2024-09-29 16:49:57.793985] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:36.359 00:11:36.359 Latency(us) 00:11:36.359 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:36.359 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:36.359 raid_bdev1 : 7.81 101.31 303.92 0.00 0.00 13507.14 273.66 113099.68 00:11:36.359 =================================================================================================================== 00:11:36.359 Total : 101.31 303.92 0.00 0.00 13507.14 273.66 113099.68 00:11:36.359 [2024-09-29 16:49:57.873153] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:36.359 [2024-09-29 16:49:57.873259] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:36.359 [2024-09-29 16:49:57.873382] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:36.359 [2024-09-29 16:49:57.873441] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:36.359 { 00:11:36.359 "results": [ 00:11:36.359 { 00:11:36.359 "job": "raid_bdev1", 00:11:36.359 "core_mask": "0x1", 00:11:36.359 "workload": "randrw", 00:11:36.359 "percentage": 50, 00:11:36.359 "status": "finished", 00:11:36.359 "queue_depth": 2, 00:11:36.359 "io_size": 3145728, 00:11:36.359 "runtime": 7.807894, 00:11:36.359 "iops": 101.30772779446032, 00:11:36.359 "mibps": 303.92318338338094, 00:11:36.359 "io_failed": 0, 00:11:36.359 "io_timeout": 0, 00:11:36.359 "avg_latency_us": 13507.138822672092, 00:11:36.359 "min_latency_us": 273.6628820960699, 00:11:36.359 "max_latency_us": 113099.68209606987 00:11:36.359 } 00:11:36.359 ], 00:11:36.359 "core_count": 1 00:11:36.359 } 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:36.359 16:49:57 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:36.619 /dev/nbd0 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:36.619 1+0 records in 00:11:36.619 1+0 records out 00:11:36.619 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000547549 s, 7.5 MB/s 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:36.619 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:36.879 /dev/nbd1 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:36.879 1+0 records in 00:11:36.879 1+0 records out 00:11:36.879 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000479576 s, 8.5 MB/s 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:36.879 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:36.880 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:36.880 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:36.880 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:36.880 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:36.880 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:36.880 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:36.880 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:37.140 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 86792 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 86792 ']' 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 86792 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 86792 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 86792' 00:11:37.399 killing process with pid 86792 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 86792 00:11:37.399 Received shutdown signal, test time was about 8.888442 seconds 00:11:37.399 00:11:37.399 Latency(us) 00:11:37.399 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:37.399 =================================================================================================================== 00:11:37.399 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:37.399 [2024-09-29 16:49:58.948173] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:37.399 16:49:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 86792 00:11:37.399 [2024-09-29 16:49:58.974016] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:37.660 00:11:37.660 real 0m10.824s 00:11:37.660 user 0m13.922s 00:11:37.660 sys 0m1.402s 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:37.660 ************************************ 00:11:37.660 END TEST raid_rebuild_test_io 00:11:37.660 ************************************ 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:11:37.660 16:49:59 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:11:37.660 16:49:59 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:37.660 16:49:59 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:37.660 16:49:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:37.660 ************************************ 00:11:37.660 START TEST raid_rebuild_test_sb_io 00:11:37.660 ************************************ 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true true true 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87154 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87154 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 87154 ']' 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:37.660 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:37.660 16:49:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:37.920 [2024-09-29 16:49:59.366003] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:11:37.920 [2024-09-29 16:49:59.366209] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87154 ] 00:11:37.920 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:37.920 Zero copy mechanism will not be used. 00:11:37.920 [2024-09-29 16:49:59.509662] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:37.920 [2024-09-29 16:49:59.555358] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:38.179 [2024-09-29 16:49:59.598326] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:38.179 [2024-09-29 16:49:59.598420] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 BaseBdev1_malloc 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 [2024-09-29 16:50:00.204614] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:38.749 [2024-09-29 16:50:00.204752] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:38.749 [2024-09-29 16:50:00.204804] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:38.749 [2024-09-29 16:50:00.204882] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:38.749 [2024-09-29 16:50:00.207018] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:38.749 [2024-09-29 16:50:00.207093] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:38.749 BaseBdev1 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 BaseBdev2_malloc 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 [2024-09-29 16:50:00.248598] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:38.749 [2024-09-29 16:50:00.248693] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:38.749 [2024-09-29 16:50:00.248762] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:38.749 [2024-09-29 16:50:00.248783] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:38.749 [2024-09-29 16:50:00.253256] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:38.749 [2024-09-29 16:50:00.253323] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:38.749 BaseBdev2 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 spare_malloc 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 spare_delay 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 [2024-09-29 16:50:00.290672] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:38.749 [2024-09-29 16:50:00.290749] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:38.749 [2024-09-29 16:50:00.290771] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:38.749 [2024-09-29 16:50:00.290778] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:38.749 [2024-09-29 16:50:00.292910] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:38.749 [2024-09-29 16:50:00.292943] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:38.749 spare 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.749 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.749 [2024-09-29 16:50:00.302697] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:38.749 [2024-09-29 16:50:00.304542] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:38.749 [2024-09-29 16:50:00.304687] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:38.750 [2024-09-29 16:50:00.304698] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:38.750 [2024-09-29 16:50:00.305001] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:38.750 [2024-09-29 16:50:00.305212] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:38.750 [2024-09-29 16:50:00.305230] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:38.750 [2024-09-29 16:50:00.305337] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:38.750 "name": "raid_bdev1", 00:11:38.750 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:38.750 "strip_size_kb": 0, 00:11:38.750 "state": "online", 00:11:38.750 "raid_level": "raid1", 00:11:38.750 "superblock": true, 00:11:38.750 "num_base_bdevs": 2, 00:11:38.750 "num_base_bdevs_discovered": 2, 00:11:38.750 "num_base_bdevs_operational": 2, 00:11:38.750 "base_bdevs_list": [ 00:11:38.750 { 00:11:38.750 "name": "BaseBdev1", 00:11:38.750 "uuid": "38ccd6e4-8ecc-58d3-baff-ee5a58f42079", 00:11:38.750 "is_configured": true, 00:11:38.750 "data_offset": 2048, 00:11:38.750 "data_size": 63488 00:11:38.750 }, 00:11:38.750 { 00:11:38.750 "name": "BaseBdev2", 00:11:38.750 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:38.750 "is_configured": true, 00:11:38.750 "data_offset": 2048, 00:11:38.750 "data_size": 63488 00:11:38.750 } 00:11:38.750 ] 00:11:38.750 }' 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:38.750 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.319 [2024-09-29 16:50:00.730167] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.319 [2024-09-29 16:50:00.829795] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:39.319 "name": "raid_bdev1", 00:11:39.319 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:39.319 "strip_size_kb": 0, 00:11:39.319 "state": "online", 00:11:39.319 "raid_level": "raid1", 00:11:39.319 "superblock": true, 00:11:39.319 "num_base_bdevs": 2, 00:11:39.319 "num_base_bdevs_discovered": 1, 00:11:39.319 "num_base_bdevs_operational": 1, 00:11:39.319 "base_bdevs_list": [ 00:11:39.319 { 00:11:39.319 "name": null, 00:11:39.319 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:39.319 "is_configured": false, 00:11:39.319 "data_offset": 0, 00:11:39.319 "data_size": 63488 00:11:39.319 }, 00:11:39.319 { 00:11:39.319 "name": "BaseBdev2", 00:11:39.319 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:39.319 "is_configured": true, 00:11:39.319 "data_offset": 2048, 00:11:39.319 "data_size": 63488 00:11:39.319 } 00:11:39.319 ] 00:11:39.319 }' 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:39.319 16:50:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.319 [2024-09-29 16:50:00.915631] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:39.319 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:39.319 Zero copy mechanism will not be used. 00:11:39.319 Running I/O for 60 seconds... 00:11:39.888 16:50:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:39.888 16:50:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.888 16:50:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:39.888 [2024-09-29 16:50:01.284948] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:39.888 16:50:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.888 16:50:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:39.888 [2024-09-29 16:50:01.331530] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:39.889 [2024-09-29 16:50:01.333528] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:39.889 [2024-09-29 16:50:01.435690] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:39.889 [2024-09-29 16:50:01.436262] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:40.148 [2024-09-29 16:50:01.670855] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:40.148 [2024-09-29 16:50:01.671248] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:40.407 206.00 IOPS, 618.00 MiB/s [2024-09-29 16:50:02.004938] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:11:40.667 [2024-09-29 16:50:02.223852] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:40.667 [2024-09-29 16:50:02.224242] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:40.667 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:40.927 "name": "raid_bdev1", 00:11:40.927 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:40.927 "strip_size_kb": 0, 00:11:40.927 "state": "online", 00:11:40.927 "raid_level": "raid1", 00:11:40.927 "superblock": true, 00:11:40.927 "num_base_bdevs": 2, 00:11:40.927 "num_base_bdevs_discovered": 2, 00:11:40.927 "num_base_bdevs_operational": 2, 00:11:40.927 "process": { 00:11:40.927 "type": "rebuild", 00:11:40.927 "target": "spare", 00:11:40.927 "progress": { 00:11:40.927 "blocks": 10240, 00:11:40.927 "percent": 16 00:11:40.927 } 00:11:40.927 }, 00:11:40.927 "base_bdevs_list": [ 00:11:40.927 { 00:11:40.927 "name": "spare", 00:11:40.927 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:40.927 "is_configured": true, 00:11:40.927 "data_offset": 2048, 00:11:40.927 "data_size": 63488 00:11:40.927 }, 00:11:40.927 { 00:11:40.927 "name": "BaseBdev2", 00:11:40.927 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:40.927 "is_configured": true, 00:11:40.927 "data_offset": 2048, 00:11:40.927 "data_size": 63488 00:11:40.927 } 00:11:40.927 ] 00:11:40.927 }' 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:40.927 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:40.927 [2024-09-29 16:50:02.457261] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:40.927 [2024-09-29 16:50:02.552996] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:40.927 [2024-09-29 16:50:02.553380] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:41.187 [2024-09-29 16:50:02.659375] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:41.187 [2024-09-29 16:50:02.666579] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.187 [2024-09-29 16:50:02.666656] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:41.187 [2024-09-29 16:50:02.666675] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:41.187 [2024-09-29 16:50:02.678652] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.187 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:41.187 "name": "raid_bdev1", 00:11:41.188 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:41.188 "strip_size_kb": 0, 00:11:41.188 "state": "online", 00:11:41.188 "raid_level": "raid1", 00:11:41.188 "superblock": true, 00:11:41.188 "num_base_bdevs": 2, 00:11:41.188 "num_base_bdevs_discovered": 1, 00:11:41.188 "num_base_bdevs_operational": 1, 00:11:41.188 "base_bdevs_list": [ 00:11:41.188 { 00:11:41.188 "name": null, 00:11:41.188 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:41.188 "is_configured": false, 00:11:41.188 "data_offset": 0, 00:11:41.188 "data_size": 63488 00:11:41.188 }, 00:11:41.188 { 00:11:41.188 "name": "BaseBdev2", 00:11:41.188 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:41.188 "is_configured": true, 00:11:41.188 "data_offset": 2048, 00:11:41.188 "data_size": 63488 00:11:41.188 } 00:11:41.188 ] 00:11:41.188 }' 00:11:41.188 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:41.188 16:50:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.447 168.50 IOPS, 505.50 MiB/s 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:41.447 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:41.447 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:41.448 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:41.448 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:41.448 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.448 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.448 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.448 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.448 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:41.708 "name": "raid_bdev1", 00:11:41.708 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:41.708 "strip_size_kb": 0, 00:11:41.708 "state": "online", 00:11:41.708 "raid_level": "raid1", 00:11:41.708 "superblock": true, 00:11:41.708 "num_base_bdevs": 2, 00:11:41.708 "num_base_bdevs_discovered": 1, 00:11:41.708 "num_base_bdevs_operational": 1, 00:11:41.708 "base_bdevs_list": [ 00:11:41.708 { 00:11:41.708 "name": null, 00:11:41.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:41.708 "is_configured": false, 00:11:41.708 "data_offset": 0, 00:11:41.708 "data_size": 63488 00:11:41.708 }, 00:11:41.708 { 00:11:41.708 "name": "BaseBdev2", 00:11:41.708 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:41.708 "is_configured": true, 00:11:41.708 "data_offset": 2048, 00:11:41.708 "data_size": 63488 00:11:41.708 } 00:11:41.708 ] 00:11:41.708 }' 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:41.708 [2024-09-29 16:50:03.237164] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.708 16:50:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:41.708 [2024-09-29 16:50:03.291023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:11:41.708 [2024-09-29 16:50:03.292981] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:41.968 [2024-09-29 16:50:03.416774] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:41.968 [2024-09-29 16:50:03.417247] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:11:41.968 [2024-09-29 16:50:03.640525] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:42.229 [2024-09-29 16:50:03.640982] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:11:42.489 178.67 IOPS, 536.00 MiB/s [2024-09-29 16:50:04.091156] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:42.489 [2024-09-29 16:50:04.091554] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:42.749 "name": "raid_bdev1", 00:11:42.749 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:42.749 "strip_size_kb": 0, 00:11:42.749 "state": "online", 00:11:42.749 "raid_level": "raid1", 00:11:42.749 "superblock": true, 00:11:42.749 "num_base_bdevs": 2, 00:11:42.749 "num_base_bdevs_discovered": 2, 00:11:42.749 "num_base_bdevs_operational": 2, 00:11:42.749 "process": { 00:11:42.749 "type": "rebuild", 00:11:42.749 "target": "spare", 00:11:42.749 "progress": { 00:11:42.749 "blocks": 10240, 00:11:42.749 "percent": 16 00:11:42.749 } 00:11:42.749 }, 00:11:42.749 "base_bdevs_list": [ 00:11:42.749 { 00:11:42.749 "name": "spare", 00:11:42.749 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:42.749 "is_configured": true, 00:11:42.749 "data_offset": 2048, 00:11:42.749 "data_size": 63488 00:11:42.749 }, 00:11:42.749 { 00:11:42.749 "name": "BaseBdev2", 00:11:42.749 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:42.749 "is_configured": true, 00:11:42.749 "data_offset": 2048, 00:11:42.749 "data_size": 63488 00:11:42.749 } 00:11:42.749 ] 00:11:42.749 }' 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:42.749 [2024-09-29 16:50:04.402115] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:42.749 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=329 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:42.749 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.009 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:43.009 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:43.009 "name": "raid_bdev1", 00:11:43.009 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:43.009 "strip_size_kb": 0, 00:11:43.009 "state": "online", 00:11:43.009 "raid_level": "raid1", 00:11:43.009 "superblock": true, 00:11:43.009 "num_base_bdevs": 2, 00:11:43.009 "num_base_bdevs_discovered": 2, 00:11:43.009 "num_base_bdevs_operational": 2, 00:11:43.009 "process": { 00:11:43.009 "type": "rebuild", 00:11:43.009 "target": "spare", 00:11:43.009 "progress": { 00:11:43.009 "blocks": 14336, 00:11:43.009 "percent": 22 00:11:43.009 } 00:11:43.009 }, 00:11:43.009 "base_bdevs_list": [ 00:11:43.010 { 00:11:43.010 "name": "spare", 00:11:43.010 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:43.010 "is_configured": true, 00:11:43.010 "data_offset": 2048, 00:11:43.010 "data_size": 63488 00:11:43.010 }, 00:11:43.010 { 00:11:43.010 "name": "BaseBdev2", 00:11:43.010 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:43.010 "is_configured": true, 00:11:43.010 "data_offset": 2048, 00:11:43.010 "data_size": 63488 00:11:43.010 } 00:11:43.010 ] 00:11:43.010 }' 00:11:43.010 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:43.010 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:43.010 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:43.010 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:43.010 16:50:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:43.010 [2024-09-29 16:50:04.611277] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:11:43.269 153.75 IOPS, 461.25 MiB/s [2024-09-29 16:50:04.929197] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:44.210 "name": "raid_bdev1", 00:11:44.210 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:44.210 "strip_size_kb": 0, 00:11:44.210 "state": "online", 00:11:44.210 "raid_level": "raid1", 00:11:44.210 "superblock": true, 00:11:44.210 "num_base_bdevs": 2, 00:11:44.210 "num_base_bdevs_discovered": 2, 00:11:44.210 "num_base_bdevs_operational": 2, 00:11:44.210 "process": { 00:11:44.210 "type": "rebuild", 00:11:44.210 "target": "spare", 00:11:44.210 "progress": { 00:11:44.210 "blocks": 30720, 00:11:44.210 "percent": 48 00:11:44.210 } 00:11:44.210 }, 00:11:44.210 "base_bdevs_list": [ 00:11:44.210 { 00:11:44.210 "name": "spare", 00:11:44.210 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:44.210 "is_configured": true, 00:11:44.210 "data_offset": 2048, 00:11:44.210 "data_size": 63488 00:11:44.210 }, 00:11:44.210 { 00:11:44.210 "name": "BaseBdev2", 00:11:44.210 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:44.210 "is_configured": true, 00:11:44.210 "data_offset": 2048, 00:11:44.210 "data_size": 63488 00:11:44.210 } 00:11:44.210 ] 00:11:44.210 }' 00:11:44.210 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:44.211 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:44.211 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:44.211 [2024-09-29 16:50:05.648327] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:11:44.211 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:44.211 16:50:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:44.211 [2024-09-29 16:50:05.862775] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:44.211 [2024-09-29 16:50:05.863077] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:11:44.730 134.00 IOPS, 402.00 MiB/s [2024-09-29 16:50:06.193855] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:11:44.730 [2024-09-29 16:50:06.194260] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.299 "name": "raid_bdev1", 00:11:45.299 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:45.299 "strip_size_kb": 0, 00:11:45.299 "state": "online", 00:11:45.299 "raid_level": "raid1", 00:11:45.299 "superblock": true, 00:11:45.299 "num_base_bdevs": 2, 00:11:45.299 "num_base_bdevs_discovered": 2, 00:11:45.299 "num_base_bdevs_operational": 2, 00:11:45.299 "process": { 00:11:45.299 "type": "rebuild", 00:11:45.299 "target": "spare", 00:11:45.299 "progress": { 00:11:45.299 "blocks": 45056, 00:11:45.299 "percent": 70 00:11:45.299 } 00:11:45.299 }, 00:11:45.299 "base_bdevs_list": [ 00:11:45.299 { 00:11:45.299 "name": "spare", 00:11:45.299 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:45.299 "is_configured": true, 00:11:45.299 "data_offset": 2048, 00:11:45.299 "data_size": 63488 00:11:45.299 }, 00:11:45.299 { 00:11:45.299 "name": "BaseBdev2", 00:11:45.299 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:45.299 "is_configured": true, 00:11:45.299 "data_offset": 2048, 00:11:45.299 "data_size": 63488 00:11:45.299 } 00:11:45.299 ] 00:11:45.299 }' 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:45.299 16:50:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:45.299 118.83 IOPS, 356.50 MiB/s [2024-09-29 16:50:06.949599] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:11:45.559 [2024-09-29 16:50:07.055540] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:11:46.128 [2024-09-29 16:50:07.596120] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:46.128 [2024-09-29 16:50:07.700985] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:46.128 [2024-09-29 16:50:07.703399] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:46.388 "name": "raid_bdev1", 00:11:46.388 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:46.388 "strip_size_kb": 0, 00:11:46.388 "state": "online", 00:11:46.388 "raid_level": "raid1", 00:11:46.388 "superblock": true, 00:11:46.388 "num_base_bdevs": 2, 00:11:46.388 "num_base_bdevs_discovered": 2, 00:11:46.388 "num_base_bdevs_operational": 2, 00:11:46.388 "base_bdevs_list": [ 00:11:46.388 { 00:11:46.388 "name": "spare", 00:11:46.388 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:46.388 "is_configured": true, 00:11:46.388 "data_offset": 2048, 00:11:46.388 "data_size": 63488 00:11:46.388 }, 00:11:46.388 { 00:11:46.388 "name": "BaseBdev2", 00:11:46.388 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:46.388 "is_configured": true, 00:11:46.388 "data_offset": 2048, 00:11:46.388 "data_size": 63488 00:11:46.388 } 00:11:46.388 ] 00:11:46.388 }' 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:46.388 106.57 IOPS, 319.71 MiB/s 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.388 16:50:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:46.388 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:46.388 "name": "raid_bdev1", 00:11:46.388 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:46.388 "strip_size_kb": 0, 00:11:46.388 "state": "online", 00:11:46.388 "raid_level": "raid1", 00:11:46.388 "superblock": true, 00:11:46.388 "num_base_bdevs": 2, 00:11:46.388 "num_base_bdevs_discovered": 2, 00:11:46.388 "num_base_bdevs_operational": 2, 00:11:46.388 "base_bdevs_list": [ 00:11:46.388 { 00:11:46.388 "name": "spare", 00:11:46.388 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:46.388 "is_configured": true, 00:11:46.389 "data_offset": 2048, 00:11:46.389 "data_size": 63488 00:11:46.389 }, 00:11:46.389 { 00:11:46.389 "name": "BaseBdev2", 00:11:46.389 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:46.389 "is_configured": true, 00:11:46.389 "data_offset": 2048, 00:11:46.389 "data_size": 63488 00:11:46.389 } 00:11:46.389 ] 00:11:46.389 }' 00:11:46.389 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:46.389 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:46.389 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:46.648 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:46.649 "name": "raid_bdev1", 00:11:46.649 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:46.649 "strip_size_kb": 0, 00:11:46.649 "state": "online", 00:11:46.649 "raid_level": "raid1", 00:11:46.649 "superblock": true, 00:11:46.649 "num_base_bdevs": 2, 00:11:46.649 "num_base_bdevs_discovered": 2, 00:11:46.649 "num_base_bdevs_operational": 2, 00:11:46.649 "base_bdevs_list": [ 00:11:46.649 { 00:11:46.649 "name": "spare", 00:11:46.649 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:46.649 "is_configured": true, 00:11:46.649 "data_offset": 2048, 00:11:46.649 "data_size": 63488 00:11:46.649 }, 00:11:46.649 { 00:11:46.649 "name": "BaseBdev2", 00:11:46.649 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:46.649 "is_configured": true, 00:11:46.649 "data_offset": 2048, 00:11:46.649 "data_size": 63488 00:11:46.649 } 00:11:46.649 ] 00:11:46.649 }' 00:11:46.649 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:46.649 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:46.909 [2024-09-29 16:50:08.482390] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:46.909 [2024-09-29 16:50:08.482477] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:46.909 00:11:46.909 Latency(us) 00:11:46.909 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:46.909 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:11:46.909 raid_bdev1 : 7.66 101.00 303.00 0.00 0.00 12850.25 266.51 112641.79 00:11:46.909 =================================================================================================================== 00:11:46.909 Total : 101.00 303.00 0.00 0.00 12850.25 266.51 112641.79 00:11:46.909 [2024-09-29 16:50:08.569329] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:46.909 [2024-09-29 16:50:08.569400] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:46.909 { 00:11:46.909 "results": [ 00:11:46.909 { 00:11:46.909 "job": "raid_bdev1", 00:11:46.909 "core_mask": "0x1", 00:11:46.909 "workload": "randrw", 00:11:46.909 "percentage": 50, 00:11:46.909 "status": "finished", 00:11:46.909 "queue_depth": 2, 00:11:46.909 "io_size": 3145728, 00:11:46.909 "runtime": 7.66325, 00:11:46.909 "iops": 101.00153329201056, 00:11:46.909 "mibps": 303.0045998760317, 00:11:46.909 "io_failed": 0, 00:11:46.909 "io_timeout": 0, 00:11:46.909 "avg_latency_us": 12850.251948139872, 00:11:46.909 "min_latency_us": 266.5082969432314, 00:11:46.909 "max_latency_us": 112641.78864628822 00:11:46.909 } 00:11:46.909 ], 00:11:46.909 "core_count": 1 00:11:46.909 } 00:11:46.909 [2024-09-29 16:50:08.569529] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:46.909 [2024-09-29 16:50:08.569546] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:11:46.909 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:47.169 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:11:47.169 /dev/nbd0 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:47.429 1+0 records in 00:11:47.429 1+0 records out 00:11:47.429 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000465046 s, 8.8 MB/s 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:47.429 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:47.430 16:50:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:11:47.430 /dev/nbd1 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:47.690 1+0 records in 00:11:47.690 1+0 records out 00:11:47.690 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00051325 s, 8.0 MB/s 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:47.690 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:47.957 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.225 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.225 [2024-09-29 16:50:09.650449] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:48.225 [2024-09-29 16:50:09.650503] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:48.225 [2024-09-29 16:50:09.650527] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:11:48.225 [2024-09-29 16:50:09.650536] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:48.225 [2024-09-29 16:50:09.652783] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:48.225 [2024-09-29 16:50:09.652817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:48.226 [2024-09-29 16:50:09.652905] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:48.226 [2024-09-29 16:50:09.652939] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:48.226 [2024-09-29 16:50:09.653071] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:48.226 spare 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.226 [2024-09-29 16:50:09.752963] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:11:48.226 [2024-09-29 16:50:09.752992] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:48.226 [2024-09-29 16:50:09.753242] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027720 00:11:48.226 [2024-09-29 16:50:09.753384] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:11:48.226 [2024-09-29 16:50:09.753394] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:11:48.226 [2024-09-29 16:50:09.753527] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:48.226 "name": "raid_bdev1", 00:11:48.226 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:48.226 "strip_size_kb": 0, 00:11:48.226 "state": "online", 00:11:48.226 "raid_level": "raid1", 00:11:48.226 "superblock": true, 00:11:48.226 "num_base_bdevs": 2, 00:11:48.226 "num_base_bdevs_discovered": 2, 00:11:48.226 "num_base_bdevs_operational": 2, 00:11:48.226 "base_bdevs_list": [ 00:11:48.226 { 00:11:48.226 "name": "spare", 00:11:48.226 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:48.226 "is_configured": true, 00:11:48.226 "data_offset": 2048, 00:11:48.226 "data_size": 63488 00:11:48.226 }, 00:11:48.226 { 00:11:48.226 "name": "BaseBdev2", 00:11:48.226 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:48.226 "is_configured": true, 00:11:48.226 "data_offset": 2048, 00:11:48.226 "data_size": 63488 00:11:48.226 } 00:11:48.226 ] 00:11:48.226 }' 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:48.226 16:50:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.828 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:48.828 "name": "raid_bdev1", 00:11:48.828 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:48.828 "strip_size_kb": 0, 00:11:48.828 "state": "online", 00:11:48.828 "raid_level": "raid1", 00:11:48.828 "superblock": true, 00:11:48.828 "num_base_bdevs": 2, 00:11:48.828 "num_base_bdevs_discovered": 2, 00:11:48.828 "num_base_bdevs_operational": 2, 00:11:48.828 "base_bdevs_list": [ 00:11:48.828 { 00:11:48.828 "name": "spare", 00:11:48.828 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:48.828 "is_configured": true, 00:11:48.828 "data_offset": 2048, 00:11:48.828 "data_size": 63488 00:11:48.828 }, 00:11:48.828 { 00:11:48.828 "name": "BaseBdev2", 00:11:48.828 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:48.828 "is_configured": true, 00:11:48.828 "data_offset": 2048, 00:11:48.829 "data_size": 63488 00:11:48.829 } 00:11:48.829 ] 00:11:48.829 }' 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.829 [2024-09-29 16:50:10.401289] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:48.829 "name": "raid_bdev1", 00:11:48.829 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:48.829 "strip_size_kb": 0, 00:11:48.829 "state": "online", 00:11:48.829 "raid_level": "raid1", 00:11:48.829 "superblock": true, 00:11:48.829 "num_base_bdevs": 2, 00:11:48.829 "num_base_bdevs_discovered": 1, 00:11:48.829 "num_base_bdevs_operational": 1, 00:11:48.829 "base_bdevs_list": [ 00:11:48.829 { 00:11:48.829 "name": null, 00:11:48.829 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:48.829 "is_configured": false, 00:11:48.829 "data_offset": 0, 00:11:48.829 "data_size": 63488 00:11:48.829 }, 00:11:48.829 { 00:11:48.829 "name": "BaseBdev2", 00:11:48.829 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:48.829 "is_configured": true, 00:11:48.829 "data_offset": 2048, 00:11:48.829 "data_size": 63488 00:11:48.829 } 00:11:48.829 ] 00:11:48.829 }' 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:48.829 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.397 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:49.397 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.397 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:49.397 [2024-09-29 16:50:10.808648] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:49.397 [2024-09-29 16:50:10.808843] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:49.397 [2024-09-29 16:50:10.808860] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:49.397 [2024-09-29 16:50:10.808905] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:49.397 [2024-09-29 16:50:10.813240] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000277f0 00:11:49.397 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.397 16:50:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:11:49.398 [2024-09-29 16:50:10.815120] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:50.337 "name": "raid_bdev1", 00:11:50.337 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:50.337 "strip_size_kb": 0, 00:11:50.337 "state": "online", 00:11:50.337 "raid_level": "raid1", 00:11:50.337 "superblock": true, 00:11:50.337 "num_base_bdevs": 2, 00:11:50.337 "num_base_bdevs_discovered": 2, 00:11:50.337 "num_base_bdevs_operational": 2, 00:11:50.337 "process": { 00:11:50.337 "type": "rebuild", 00:11:50.337 "target": "spare", 00:11:50.337 "progress": { 00:11:50.337 "blocks": 20480, 00:11:50.337 "percent": 32 00:11:50.337 } 00:11:50.337 }, 00:11:50.337 "base_bdevs_list": [ 00:11:50.337 { 00:11:50.337 "name": "spare", 00:11:50.337 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:50.337 "is_configured": true, 00:11:50.337 "data_offset": 2048, 00:11:50.337 "data_size": 63488 00:11:50.337 }, 00:11:50.337 { 00:11:50.337 "name": "BaseBdev2", 00:11:50.337 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:50.337 "is_configured": true, 00:11:50.337 "data_offset": 2048, 00:11:50.337 "data_size": 63488 00:11:50.337 } 00:11:50.337 ] 00:11:50.337 }' 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:50.337 16:50:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.337 [2024-09-29 16:50:11.983334] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:50.597 [2024-09-29 16:50:12.019184] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:50.597 [2024-09-29 16:50:12.019308] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:50.597 [2024-09-29 16:50:12.019364] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:50.597 [2024-09-29 16:50:12.019411] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:50.597 "name": "raid_bdev1", 00:11:50.597 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:50.597 "strip_size_kb": 0, 00:11:50.597 "state": "online", 00:11:50.597 "raid_level": "raid1", 00:11:50.597 "superblock": true, 00:11:50.597 "num_base_bdevs": 2, 00:11:50.597 "num_base_bdevs_discovered": 1, 00:11:50.597 "num_base_bdevs_operational": 1, 00:11:50.597 "base_bdevs_list": [ 00:11:50.597 { 00:11:50.597 "name": null, 00:11:50.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:50.597 "is_configured": false, 00:11:50.597 "data_offset": 0, 00:11:50.597 "data_size": 63488 00:11:50.597 }, 00:11:50.597 { 00:11:50.597 "name": "BaseBdev2", 00:11:50.597 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:50.597 "is_configured": true, 00:11:50.597 "data_offset": 2048, 00:11:50.597 "data_size": 63488 00:11:50.597 } 00:11:50.597 ] 00:11:50.597 }' 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:50.597 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.857 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:50.857 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:50.857 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:50.857 [2024-09-29 16:50:12.507284] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:50.857 [2024-09-29 16:50:12.507399] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:50.857 [2024-09-29 16:50:12.507442] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:50.857 [2024-09-29 16:50:12.507473] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:50.857 [2024-09-29 16:50:12.507953] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:50.857 [2024-09-29 16:50:12.508016] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:50.857 [2024-09-29 16:50:12.508146] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:11:50.857 [2024-09-29 16:50:12.508166] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:11:50.857 [2024-09-29 16:50:12.508176] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:11:50.857 [2024-09-29 16:50:12.508201] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:50.857 spare 00:11:50.857 [2024-09-29 16:50:12.512399] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000278c0 00:11:50.857 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:50.857 16:50:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:11:50.857 [2024-09-29 16:50:12.514227] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:52.238 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:52.238 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:52.238 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:52.238 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:52.238 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:52.238 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.238 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:52.239 "name": "raid_bdev1", 00:11:52.239 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:52.239 "strip_size_kb": 0, 00:11:52.239 "state": "online", 00:11:52.239 "raid_level": "raid1", 00:11:52.239 "superblock": true, 00:11:52.239 "num_base_bdevs": 2, 00:11:52.239 "num_base_bdevs_discovered": 2, 00:11:52.239 "num_base_bdevs_operational": 2, 00:11:52.239 "process": { 00:11:52.239 "type": "rebuild", 00:11:52.239 "target": "spare", 00:11:52.239 "progress": { 00:11:52.239 "blocks": 20480, 00:11:52.239 "percent": 32 00:11:52.239 } 00:11:52.239 }, 00:11:52.239 "base_bdevs_list": [ 00:11:52.239 { 00:11:52.239 "name": "spare", 00:11:52.239 "uuid": "7fbc0801-39ca-56c3-abf3-67882bfd7bb3", 00:11:52.239 "is_configured": true, 00:11:52.239 "data_offset": 2048, 00:11:52.239 "data_size": 63488 00:11:52.239 }, 00:11:52.239 { 00:11:52.239 "name": "BaseBdev2", 00:11:52.239 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:52.239 "is_configured": true, 00:11:52.239 "data_offset": 2048, 00:11:52.239 "data_size": 63488 00:11:52.239 } 00:11:52.239 ] 00:11:52.239 }' 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.239 [2024-09-29 16:50:13.662550] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:52.239 [2024-09-29 16:50:13.718390] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:52.239 [2024-09-29 16:50:13.718462] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:52.239 [2024-09-29 16:50:13.718478] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:52.239 [2024-09-29 16:50:13.718485] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:52.239 "name": "raid_bdev1", 00:11:52.239 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:52.239 "strip_size_kb": 0, 00:11:52.239 "state": "online", 00:11:52.239 "raid_level": "raid1", 00:11:52.239 "superblock": true, 00:11:52.239 "num_base_bdevs": 2, 00:11:52.239 "num_base_bdevs_discovered": 1, 00:11:52.239 "num_base_bdevs_operational": 1, 00:11:52.239 "base_bdevs_list": [ 00:11:52.239 { 00:11:52.239 "name": null, 00:11:52.239 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.239 "is_configured": false, 00:11:52.239 "data_offset": 0, 00:11:52.239 "data_size": 63488 00:11:52.239 }, 00:11:52.239 { 00:11:52.239 "name": "BaseBdev2", 00:11:52.239 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:52.239 "is_configured": true, 00:11:52.239 "data_offset": 2048, 00:11:52.239 "data_size": 63488 00:11:52.239 } 00:11:52.239 ] 00:11:52.239 }' 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:52.239 16:50:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.498 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:52.498 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:52.498 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:52.498 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:52.498 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:52.758 "name": "raid_bdev1", 00:11:52.758 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:52.758 "strip_size_kb": 0, 00:11:52.758 "state": "online", 00:11:52.758 "raid_level": "raid1", 00:11:52.758 "superblock": true, 00:11:52.758 "num_base_bdevs": 2, 00:11:52.758 "num_base_bdevs_discovered": 1, 00:11:52.758 "num_base_bdevs_operational": 1, 00:11:52.758 "base_bdevs_list": [ 00:11:52.758 { 00:11:52.758 "name": null, 00:11:52.758 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:52.758 "is_configured": false, 00:11:52.758 "data_offset": 0, 00:11:52.758 "data_size": 63488 00:11:52.758 }, 00:11:52.758 { 00:11:52.758 "name": "BaseBdev2", 00:11:52.758 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:52.758 "is_configured": true, 00:11:52.758 "data_offset": 2048, 00:11:52.758 "data_size": 63488 00:11:52.758 } 00:11:52.758 ] 00:11:52.758 }' 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:52.758 [2024-09-29 16:50:14.305999] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:52.758 [2024-09-29 16:50:14.306052] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:52.758 [2024-09-29 16:50:14.306097] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:11:52.758 [2024-09-29 16:50:14.306106] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:52.758 [2024-09-29 16:50:14.306497] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:52.758 [2024-09-29 16:50:14.306513] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:52.758 [2024-09-29 16:50:14.306584] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:11:52.758 [2024-09-29 16:50:14.306597] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:52.758 [2024-09-29 16:50:14.306606] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:52.758 [2024-09-29 16:50:14.306615] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:11:52.758 BaseBdev1 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:52.758 16:50:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.696 "name": "raid_bdev1", 00:11:53.696 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:53.696 "strip_size_kb": 0, 00:11:53.696 "state": "online", 00:11:53.696 "raid_level": "raid1", 00:11:53.696 "superblock": true, 00:11:53.696 "num_base_bdevs": 2, 00:11:53.696 "num_base_bdevs_discovered": 1, 00:11:53.696 "num_base_bdevs_operational": 1, 00:11:53.696 "base_bdevs_list": [ 00:11:53.696 { 00:11:53.696 "name": null, 00:11:53.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.696 "is_configured": false, 00:11:53.696 "data_offset": 0, 00:11:53.696 "data_size": 63488 00:11:53.696 }, 00:11:53.696 { 00:11:53.696 "name": "BaseBdev2", 00:11:53.696 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:53.696 "is_configured": true, 00:11:53.696 "data_offset": 2048, 00:11:53.696 "data_size": 63488 00:11:53.696 } 00:11:53.696 ] 00:11:53.696 }' 00:11:53.696 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.956 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:54.215 "name": "raid_bdev1", 00:11:54.215 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:54.215 "strip_size_kb": 0, 00:11:54.215 "state": "online", 00:11:54.215 "raid_level": "raid1", 00:11:54.215 "superblock": true, 00:11:54.215 "num_base_bdevs": 2, 00:11:54.215 "num_base_bdevs_discovered": 1, 00:11:54.215 "num_base_bdevs_operational": 1, 00:11:54.215 "base_bdevs_list": [ 00:11:54.215 { 00:11:54.215 "name": null, 00:11:54.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:54.215 "is_configured": false, 00:11:54.215 "data_offset": 0, 00:11:54.215 "data_size": 63488 00:11:54.215 }, 00:11:54.215 { 00:11:54.215 "name": "BaseBdev2", 00:11:54.215 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:54.215 "is_configured": true, 00:11:54.215 "data_offset": 2048, 00:11:54.215 "data_size": 63488 00:11:54.215 } 00:11:54.215 ] 00:11:54.215 }' 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:54.215 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:54.475 [2024-09-29 16:50:15.923478] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:54.475 [2024-09-29 16:50:15.923629] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:11:54.475 [2024-09-29 16:50:15.923647] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:11:54.475 request: 00:11:54.475 { 00:11:54.475 "base_bdev": "BaseBdev1", 00:11:54.475 "raid_bdev": "raid_bdev1", 00:11:54.475 "method": "bdev_raid_add_base_bdev", 00:11:54.475 "req_id": 1 00:11:54.475 } 00:11:54.475 Got JSON-RPC error response 00:11:54.475 response: 00:11:54.475 { 00:11:54.475 "code": -22, 00:11:54.475 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:11:54.475 } 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:54.475 16:50:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:55.414 "name": "raid_bdev1", 00:11:55.414 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:55.414 "strip_size_kb": 0, 00:11:55.414 "state": "online", 00:11:55.414 "raid_level": "raid1", 00:11:55.414 "superblock": true, 00:11:55.414 "num_base_bdevs": 2, 00:11:55.414 "num_base_bdevs_discovered": 1, 00:11:55.414 "num_base_bdevs_operational": 1, 00:11:55.414 "base_bdevs_list": [ 00:11:55.414 { 00:11:55.414 "name": null, 00:11:55.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.414 "is_configured": false, 00:11:55.414 "data_offset": 0, 00:11:55.414 "data_size": 63488 00:11:55.414 }, 00:11:55.414 { 00:11:55.414 "name": "BaseBdev2", 00:11:55.414 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:55.414 "is_configured": true, 00:11:55.414 "data_offset": 2048, 00:11:55.414 "data_size": 63488 00:11:55.414 } 00:11:55.414 ] 00:11:55.414 }' 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:55.414 16:50:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.984 "name": "raid_bdev1", 00:11:55.984 "uuid": "3412034c-3571-47cb-ba61-135f06dba531", 00:11:55.984 "strip_size_kb": 0, 00:11:55.984 "state": "online", 00:11:55.984 "raid_level": "raid1", 00:11:55.984 "superblock": true, 00:11:55.984 "num_base_bdevs": 2, 00:11:55.984 "num_base_bdevs_discovered": 1, 00:11:55.984 "num_base_bdevs_operational": 1, 00:11:55.984 "base_bdevs_list": [ 00:11:55.984 { 00:11:55.984 "name": null, 00:11:55.984 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.984 "is_configured": false, 00:11:55.984 "data_offset": 0, 00:11:55.984 "data_size": 63488 00:11:55.984 }, 00:11:55.984 { 00:11:55.984 "name": "BaseBdev2", 00:11:55.984 "uuid": "72873ce5-2cf3-5405-b559-2f3cb24cf3a6", 00:11:55.984 "is_configured": true, 00:11:55.984 "data_offset": 2048, 00:11:55.984 "data_size": 63488 00:11:55.984 } 00:11:55.984 ] 00:11:55.984 }' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87154 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 87154 ']' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 87154 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87154 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:55.984 killing process with pid 87154 00:11:55.984 Received shutdown signal, test time was about 16.647549 seconds 00:11:55.984 00:11:55.984 Latency(us) 00:11:55.984 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:55.984 =================================================================================================================== 00:11:55.984 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87154' 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 87154 00:11:55.984 [2024-09-29 16:50:17.533192] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:55.984 [2024-09-29 16:50:17.533330] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:55.984 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 87154 00:11:55.984 [2024-09-29 16:50:17.533385] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:55.984 [2024-09-29 16:50:17.533400] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:11:55.984 [2024-09-29 16:50:17.558994] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:11:56.245 00:11:56.245 real 0m18.513s 00:11:56.245 user 0m24.581s 00:11:56.245 sys 0m2.039s 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:11:56.245 ************************************ 00:11:56.245 END TEST raid_rebuild_test_sb_io 00:11:56.245 ************************************ 00:11:56.245 16:50:17 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:11:56.245 16:50:17 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:11:56.245 16:50:17 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:56.245 16:50:17 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:56.245 16:50:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:56.245 ************************************ 00:11:56.245 START TEST raid_rebuild_test 00:11:56.245 ************************************ 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false false true 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=87831 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 87831 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 87831 ']' 00:11:56.245 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:56.245 16:50:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:56.505 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:56.505 Zero copy mechanism will not be used. 00:11:56.505 [2024-09-29 16:50:17.955404] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:11:56.505 [2024-09-29 16:50:17.955526] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87831 ] 00:11:56.505 [2024-09-29 16:50:18.099109] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:56.505 [2024-09-29 16:50:18.144726] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:56.764 [2024-09-29 16:50:18.186802] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:56.764 [2024-09-29 16:50:18.186839] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 BaseBdev1_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 [2024-09-29 16:50:18.792765] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:57.343 [2024-09-29 16:50:18.792825] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:57.343 [2024-09-29 16:50:18.792869] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:57.343 [2024-09-29 16:50:18.792882] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:57.343 [2024-09-29 16:50:18.794999] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:57.343 [2024-09-29 16:50:18.795036] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:57.343 BaseBdev1 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 BaseBdev2_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 [2024-09-29 16:50:18.836736] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:57.343 [2024-09-29 16:50:18.836826] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:57.343 [2024-09-29 16:50:18.836874] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:57.343 [2024-09-29 16:50:18.836895] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:57.343 [2024-09-29 16:50:18.841435] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:57.343 [2024-09-29 16:50:18.841505] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:57.343 BaseBdev2 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 BaseBdev3_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 [2024-09-29 16:50:18.866924] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:11:57.343 [2024-09-29 16:50:18.867047] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:57.343 [2024-09-29 16:50:18.867076] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:57.343 [2024-09-29 16:50:18.867085] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:57.343 [2024-09-29 16:50:18.869114] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:57.343 [2024-09-29 16:50:18.869146] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:57.343 BaseBdev3 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 BaseBdev4_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 [2024-09-29 16:50:18.895368] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:11:57.343 [2024-09-29 16:50:18.895416] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:57.343 [2024-09-29 16:50:18.895454] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:57.343 [2024-09-29 16:50:18.895463] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:57.343 [2024-09-29 16:50:18.897489] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:57.343 [2024-09-29 16:50:18.897524] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:57.343 BaseBdev4 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.343 spare_malloc 00:11:57.343 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.344 spare_delay 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.344 [2024-09-29 16:50:18.935744] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:57.344 [2024-09-29 16:50:18.935786] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:57.344 [2024-09-29 16:50:18.935823] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:11:57.344 [2024-09-29 16:50:18.935831] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:57.344 [2024-09-29 16:50:18.937873] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:57.344 [2024-09-29 16:50:18.937906] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:57.344 spare 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.344 [2024-09-29 16:50:18.947799] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:57.344 [2024-09-29 16:50:18.949592] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:57.344 [2024-09-29 16:50:18.949655] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:57.344 [2024-09-29 16:50:18.949701] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:57.344 [2024-09-29 16:50:18.949785] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:57.344 [2024-09-29 16:50:18.949794] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:57.344 [2024-09-29 16:50:18.950105] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:57.344 [2024-09-29 16:50:18.950233] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:57.344 [2024-09-29 16:50:18.950246] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:57.344 [2024-09-29 16:50:18.950372] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.344 16:50:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.344 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:57.344 "name": "raid_bdev1", 00:11:57.344 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:11:57.344 "strip_size_kb": 0, 00:11:57.344 "state": "online", 00:11:57.344 "raid_level": "raid1", 00:11:57.344 "superblock": false, 00:11:57.344 "num_base_bdevs": 4, 00:11:57.344 "num_base_bdevs_discovered": 4, 00:11:57.344 "num_base_bdevs_operational": 4, 00:11:57.344 "base_bdevs_list": [ 00:11:57.344 { 00:11:57.344 "name": "BaseBdev1", 00:11:57.344 "uuid": "ce6798ae-2046-5f0e-bfab-4be15d95f742", 00:11:57.344 "is_configured": true, 00:11:57.344 "data_offset": 0, 00:11:57.344 "data_size": 65536 00:11:57.344 }, 00:11:57.344 { 00:11:57.344 "name": "BaseBdev2", 00:11:57.344 "uuid": "264c5a03-07a1-5729-bd00-296e45e7f1fa", 00:11:57.344 "is_configured": true, 00:11:57.344 "data_offset": 0, 00:11:57.344 "data_size": 65536 00:11:57.344 }, 00:11:57.344 { 00:11:57.344 "name": "BaseBdev3", 00:11:57.344 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:11:57.344 "is_configured": true, 00:11:57.344 "data_offset": 0, 00:11:57.344 "data_size": 65536 00:11:57.344 }, 00:11:57.344 { 00:11:57.344 "name": "BaseBdev4", 00:11:57.344 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:11:57.344 "is_configured": true, 00:11:57.344 "data_offset": 0, 00:11:57.344 "data_size": 65536 00:11:57.344 } 00:11:57.344 ] 00:11:57.344 }' 00:11:57.344 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:57.344 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.914 [2024-09-29 16:50:19.399370] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:57.914 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:58.173 [2024-09-29 16:50:19.666631] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:58.173 /dev/nbd0 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:58.173 1+0 records in 00:11:58.173 1+0 records out 00:11:58.173 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000242403 s, 16.9 MB/s 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:58.173 16:50:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:03.444 65536+0 records in 00:12:03.444 65536+0 records out 00:12:03.444 33554432 bytes (34 MB, 32 MiB) copied, 4.98263 s, 6.7 MB/s 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:03.444 [2024-09-29 16:50:24.923437] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:03.444 [2024-09-29 16:50:24.940620] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.444 "name": "raid_bdev1", 00:12:03.444 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:03.444 "strip_size_kb": 0, 00:12:03.444 "state": "online", 00:12:03.444 "raid_level": "raid1", 00:12:03.444 "superblock": false, 00:12:03.444 "num_base_bdevs": 4, 00:12:03.444 "num_base_bdevs_discovered": 3, 00:12:03.444 "num_base_bdevs_operational": 3, 00:12:03.444 "base_bdevs_list": [ 00:12:03.444 { 00:12:03.444 "name": null, 00:12:03.444 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.444 "is_configured": false, 00:12:03.444 "data_offset": 0, 00:12:03.444 "data_size": 65536 00:12:03.444 }, 00:12:03.444 { 00:12:03.444 "name": "BaseBdev2", 00:12:03.444 "uuid": "264c5a03-07a1-5729-bd00-296e45e7f1fa", 00:12:03.444 "is_configured": true, 00:12:03.444 "data_offset": 0, 00:12:03.444 "data_size": 65536 00:12:03.444 }, 00:12:03.444 { 00:12:03.444 "name": "BaseBdev3", 00:12:03.444 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:03.444 "is_configured": true, 00:12:03.444 "data_offset": 0, 00:12:03.444 "data_size": 65536 00:12:03.444 }, 00:12:03.444 { 00:12:03.444 "name": "BaseBdev4", 00:12:03.444 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:03.444 "is_configured": true, 00:12:03.444 "data_offset": 0, 00:12:03.444 "data_size": 65536 00:12:03.444 } 00:12:03.444 ] 00:12:03.444 }' 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.444 16:50:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:04.010 16:50:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:04.010 16:50:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.010 16:50:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:04.010 [2024-09-29 16:50:25.403857] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:04.011 [2024-09-29 16:50:25.407303] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d063c0 00:12:04.011 [2024-09-29 16:50:25.409220] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:04.011 16:50:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:04.011 16:50:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:04.945 "name": "raid_bdev1", 00:12:04.945 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:04.945 "strip_size_kb": 0, 00:12:04.945 "state": "online", 00:12:04.945 "raid_level": "raid1", 00:12:04.945 "superblock": false, 00:12:04.945 "num_base_bdevs": 4, 00:12:04.945 "num_base_bdevs_discovered": 4, 00:12:04.945 "num_base_bdevs_operational": 4, 00:12:04.945 "process": { 00:12:04.945 "type": "rebuild", 00:12:04.945 "target": "spare", 00:12:04.945 "progress": { 00:12:04.945 "blocks": 20480, 00:12:04.945 "percent": 31 00:12:04.945 } 00:12:04.945 }, 00:12:04.945 "base_bdevs_list": [ 00:12:04.945 { 00:12:04.945 "name": "spare", 00:12:04.945 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:04.945 "is_configured": true, 00:12:04.945 "data_offset": 0, 00:12:04.945 "data_size": 65536 00:12:04.945 }, 00:12:04.945 { 00:12:04.945 "name": "BaseBdev2", 00:12:04.945 "uuid": "264c5a03-07a1-5729-bd00-296e45e7f1fa", 00:12:04.945 "is_configured": true, 00:12:04.945 "data_offset": 0, 00:12:04.945 "data_size": 65536 00:12:04.945 }, 00:12:04.945 { 00:12:04.945 "name": "BaseBdev3", 00:12:04.945 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:04.945 "is_configured": true, 00:12:04.945 "data_offset": 0, 00:12:04.945 "data_size": 65536 00:12:04.945 }, 00:12:04.945 { 00:12:04.945 "name": "BaseBdev4", 00:12:04.945 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:04.945 "is_configured": true, 00:12:04.945 "data_offset": 0, 00:12:04.945 "data_size": 65536 00:12:04.945 } 00:12:04.945 ] 00:12:04.945 }' 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:04.945 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:04.945 [2024-09-29 16:50:26.559993] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:04.945 [2024-09-29 16:50:26.614131] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:04.945 [2024-09-29 16:50:26.614238] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:04.945 [2024-09-29 16:50:26.614279] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:04.945 [2024-09-29 16:50:26.614310] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:05.204 "name": "raid_bdev1", 00:12:05.204 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:05.204 "strip_size_kb": 0, 00:12:05.204 "state": "online", 00:12:05.204 "raid_level": "raid1", 00:12:05.204 "superblock": false, 00:12:05.204 "num_base_bdevs": 4, 00:12:05.204 "num_base_bdevs_discovered": 3, 00:12:05.204 "num_base_bdevs_operational": 3, 00:12:05.204 "base_bdevs_list": [ 00:12:05.204 { 00:12:05.204 "name": null, 00:12:05.204 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.204 "is_configured": false, 00:12:05.204 "data_offset": 0, 00:12:05.204 "data_size": 65536 00:12:05.204 }, 00:12:05.204 { 00:12:05.204 "name": "BaseBdev2", 00:12:05.204 "uuid": "264c5a03-07a1-5729-bd00-296e45e7f1fa", 00:12:05.204 "is_configured": true, 00:12:05.204 "data_offset": 0, 00:12:05.204 "data_size": 65536 00:12:05.204 }, 00:12:05.204 { 00:12:05.204 "name": "BaseBdev3", 00:12:05.204 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:05.204 "is_configured": true, 00:12:05.204 "data_offset": 0, 00:12:05.204 "data_size": 65536 00:12:05.204 }, 00:12:05.204 { 00:12:05.204 "name": "BaseBdev4", 00:12:05.204 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:05.204 "is_configured": true, 00:12:05.204 "data_offset": 0, 00:12:05.204 "data_size": 65536 00:12:05.204 } 00:12:05.204 ] 00:12:05.204 }' 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:05.204 16:50:26 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.470 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:05.470 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:05.470 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:05.470 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:05.471 "name": "raid_bdev1", 00:12:05.471 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:05.471 "strip_size_kb": 0, 00:12:05.471 "state": "online", 00:12:05.471 "raid_level": "raid1", 00:12:05.471 "superblock": false, 00:12:05.471 "num_base_bdevs": 4, 00:12:05.471 "num_base_bdevs_discovered": 3, 00:12:05.471 "num_base_bdevs_operational": 3, 00:12:05.471 "base_bdevs_list": [ 00:12:05.471 { 00:12:05.471 "name": null, 00:12:05.471 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.471 "is_configured": false, 00:12:05.471 "data_offset": 0, 00:12:05.471 "data_size": 65536 00:12:05.471 }, 00:12:05.471 { 00:12:05.471 "name": "BaseBdev2", 00:12:05.471 "uuid": "264c5a03-07a1-5729-bd00-296e45e7f1fa", 00:12:05.471 "is_configured": true, 00:12:05.471 "data_offset": 0, 00:12:05.471 "data_size": 65536 00:12:05.471 }, 00:12:05.471 { 00:12:05.471 "name": "BaseBdev3", 00:12:05.471 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:05.471 "is_configured": true, 00:12:05.471 "data_offset": 0, 00:12:05.471 "data_size": 65536 00:12:05.471 }, 00:12:05.471 { 00:12:05.471 "name": "BaseBdev4", 00:12:05.471 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:05.471 "is_configured": true, 00:12:05.471 "data_offset": 0, 00:12:05.471 "data_size": 65536 00:12:05.471 } 00:12:05.471 ] 00:12:05.471 }' 00:12:05.471 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:05.759 [2024-09-29 16:50:27.229532] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:05.759 [2024-09-29 16:50:27.232989] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06490 00:12:05.759 [2024-09-29 16:50:27.234886] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.759 16:50:27 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:06.717 "name": "raid_bdev1", 00:12:06.717 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:06.717 "strip_size_kb": 0, 00:12:06.717 "state": "online", 00:12:06.717 "raid_level": "raid1", 00:12:06.717 "superblock": false, 00:12:06.717 "num_base_bdevs": 4, 00:12:06.717 "num_base_bdevs_discovered": 4, 00:12:06.717 "num_base_bdevs_operational": 4, 00:12:06.717 "process": { 00:12:06.717 "type": "rebuild", 00:12:06.717 "target": "spare", 00:12:06.717 "progress": { 00:12:06.717 "blocks": 20480, 00:12:06.717 "percent": 31 00:12:06.717 } 00:12:06.717 }, 00:12:06.717 "base_bdevs_list": [ 00:12:06.717 { 00:12:06.717 "name": "spare", 00:12:06.717 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:06.717 "is_configured": true, 00:12:06.717 "data_offset": 0, 00:12:06.717 "data_size": 65536 00:12:06.717 }, 00:12:06.717 { 00:12:06.717 "name": "BaseBdev2", 00:12:06.717 "uuid": "264c5a03-07a1-5729-bd00-296e45e7f1fa", 00:12:06.717 "is_configured": true, 00:12:06.717 "data_offset": 0, 00:12:06.717 "data_size": 65536 00:12:06.717 }, 00:12:06.717 { 00:12:06.717 "name": "BaseBdev3", 00:12:06.717 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:06.717 "is_configured": true, 00:12:06.717 "data_offset": 0, 00:12:06.717 "data_size": 65536 00:12:06.717 }, 00:12:06.717 { 00:12:06.717 "name": "BaseBdev4", 00:12:06.717 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:06.717 "is_configured": true, 00:12:06.717 "data_offset": 0, 00:12:06.717 "data_size": 65536 00:12:06.717 } 00:12:06.717 ] 00:12:06.717 }' 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:06.717 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.977 [2024-09-29 16:50:28.397685] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:06.977 [2024-09-29 16:50:28.439074] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d06490 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:06.977 "name": "raid_bdev1", 00:12:06.977 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:06.977 "strip_size_kb": 0, 00:12:06.977 "state": "online", 00:12:06.977 "raid_level": "raid1", 00:12:06.977 "superblock": false, 00:12:06.977 "num_base_bdevs": 4, 00:12:06.977 "num_base_bdevs_discovered": 3, 00:12:06.977 "num_base_bdevs_operational": 3, 00:12:06.977 "process": { 00:12:06.977 "type": "rebuild", 00:12:06.977 "target": "spare", 00:12:06.977 "progress": { 00:12:06.977 "blocks": 24576, 00:12:06.977 "percent": 37 00:12:06.977 } 00:12:06.977 }, 00:12:06.977 "base_bdevs_list": [ 00:12:06.977 { 00:12:06.977 "name": "spare", 00:12:06.977 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:06.977 "is_configured": true, 00:12:06.977 "data_offset": 0, 00:12:06.977 "data_size": 65536 00:12:06.977 }, 00:12:06.977 { 00:12:06.977 "name": null, 00:12:06.977 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:06.977 "is_configured": false, 00:12:06.977 "data_offset": 0, 00:12:06.977 "data_size": 65536 00:12:06.977 }, 00:12:06.977 { 00:12:06.977 "name": "BaseBdev3", 00:12:06.977 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:06.977 "is_configured": true, 00:12:06.977 "data_offset": 0, 00:12:06.977 "data_size": 65536 00:12:06.977 }, 00:12:06.977 { 00:12:06.977 "name": "BaseBdev4", 00:12:06.977 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:06.977 "is_configured": true, 00:12:06.977 "data_offset": 0, 00:12:06.977 "data_size": 65536 00:12:06.977 } 00:12:06.977 ] 00:12:06.977 }' 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=353 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:06.977 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:06.978 "name": "raid_bdev1", 00:12:06.978 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:06.978 "strip_size_kb": 0, 00:12:06.978 "state": "online", 00:12:06.978 "raid_level": "raid1", 00:12:06.978 "superblock": false, 00:12:06.978 "num_base_bdevs": 4, 00:12:06.978 "num_base_bdevs_discovered": 3, 00:12:06.978 "num_base_bdevs_operational": 3, 00:12:06.978 "process": { 00:12:06.978 "type": "rebuild", 00:12:06.978 "target": "spare", 00:12:06.978 "progress": { 00:12:06.978 "blocks": 26624, 00:12:06.978 "percent": 40 00:12:06.978 } 00:12:06.978 }, 00:12:06.978 "base_bdevs_list": [ 00:12:06.978 { 00:12:06.978 "name": "spare", 00:12:06.978 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:06.978 "is_configured": true, 00:12:06.978 "data_offset": 0, 00:12:06.978 "data_size": 65536 00:12:06.978 }, 00:12:06.978 { 00:12:06.978 "name": null, 00:12:06.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:06.978 "is_configured": false, 00:12:06.978 "data_offset": 0, 00:12:06.978 "data_size": 65536 00:12:06.978 }, 00:12:06.978 { 00:12:06.978 "name": "BaseBdev3", 00:12:06.978 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:06.978 "is_configured": true, 00:12:06.978 "data_offset": 0, 00:12:06.978 "data_size": 65536 00:12:06.978 }, 00:12:06.978 { 00:12:06.978 "name": "BaseBdev4", 00:12:06.978 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:06.978 "is_configured": true, 00:12:06.978 "data_offset": 0, 00:12:06.978 "data_size": 65536 00:12:06.978 } 00:12:06.978 ] 00:12:06.978 }' 00:12:06.978 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:07.238 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:07.238 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:07.238 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:07.238 16:50:28 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:08.177 "name": "raid_bdev1", 00:12:08.177 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:08.177 "strip_size_kb": 0, 00:12:08.177 "state": "online", 00:12:08.177 "raid_level": "raid1", 00:12:08.177 "superblock": false, 00:12:08.177 "num_base_bdevs": 4, 00:12:08.177 "num_base_bdevs_discovered": 3, 00:12:08.177 "num_base_bdevs_operational": 3, 00:12:08.177 "process": { 00:12:08.177 "type": "rebuild", 00:12:08.177 "target": "spare", 00:12:08.177 "progress": { 00:12:08.177 "blocks": 49152, 00:12:08.177 "percent": 75 00:12:08.177 } 00:12:08.177 }, 00:12:08.177 "base_bdevs_list": [ 00:12:08.177 { 00:12:08.177 "name": "spare", 00:12:08.177 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:08.177 "is_configured": true, 00:12:08.177 "data_offset": 0, 00:12:08.177 "data_size": 65536 00:12:08.177 }, 00:12:08.177 { 00:12:08.177 "name": null, 00:12:08.177 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:08.177 "is_configured": false, 00:12:08.177 "data_offset": 0, 00:12:08.177 "data_size": 65536 00:12:08.177 }, 00:12:08.177 { 00:12:08.177 "name": "BaseBdev3", 00:12:08.177 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:08.177 "is_configured": true, 00:12:08.177 "data_offset": 0, 00:12:08.177 "data_size": 65536 00:12:08.177 }, 00:12:08.177 { 00:12:08.177 "name": "BaseBdev4", 00:12:08.177 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:08.177 "is_configured": true, 00:12:08.177 "data_offset": 0, 00:12:08.177 "data_size": 65536 00:12:08.177 } 00:12:08.177 ] 00:12:08.177 }' 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:08.177 16:50:29 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:09.116 [2024-09-29 16:50:30.446116] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:09.116 [2024-09-29 16:50:30.446229] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:09.116 [2024-09-29 16:50:30.446290] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:09.374 "name": "raid_bdev1", 00:12:09.374 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:09.374 "strip_size_kb": 0, 00:12:09.374 "state": "online", 00:12:09.374 "raid_level": "raid1", 00:12:09.374 "superblock": false, 00:12:09.374 "num_base_bdevs": 4, 00:12:09.374 "num_base_bdevs_discovered": 3, 00:12:09.374 "num_base_bdevs_operational": 3, 00:12:09.374 "base_bdevs_list": [ 00:12:09.374 { 00:12:09.374 "name": "spare", 00:12:09.374 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:09.374 "is_configured": true, 00:12:09.374 "data_offset": 0, 00:12:09.374 "data_size": 65536 00:12:09.374 }, 00:12:09.374 { 00:12:09.374 "name": null, 00:12:09.374 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.374 "is_configured": false, 00:12:09.374 "data_offset": 0, 00:12:09.374 "data_size": 65536 00:12:09.374 }, 00:12:09.374 { 00:12:09.374 "name": "BaseBdev3", 00:12:09.374 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:09.374 "is_configured": true, 00:12:09.374 "data_offset": 0, 00:12:09.374 "data_size": 65536 00:12:09.374 }, 00:12:09.374 { 00:12:09.374 "name": "BaseBdev4", 00:12:09.374 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:09.374 "is_configured": true, 00:12:09.374 "data_offset": 0, 00:12:09.374 "data_size": 65536 00:12:09.374 } 00:12:09.374 ] 00:12:09.374 }' 00:12:09.374 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.375 16:50:30 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.375 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.375 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:09.375 "name": "raid_bdev1", 00:12:09.375 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:09.375 "strip_size_kb": 0, 00:12:09.375 "state": "online", 00:12:09.375 "raid_level": "raid1", 00:12:09.375 "superblock": false, 00:12:09.375 "num_base_bdevs": 4, 00:12:09.375 "num_base_bdevs_discovered": 3, 00:12:09.375 "num_base_bdevs_operational": 3, 00:12:09.375 "base_bdevs_list": [ 00:12:09.375 { 00:12:09.375 "name": "spare", 00:12:09.375 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:09.375 "is_configured": true, 00:12:09.375 "data_offset": 0, 00:12:09.375 "data_size": 65536 00:12:09.375 }, 00:12:09.375 { 00:12:09.375 "name": null, 00:12:09.375 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.375 "is_configured": false, 00:12:09.375 "data_offset": 0, 00:12:09.375 "data_size": 65536 00:12:09.375 }, 00:12:09.375 { 00:12:09.375 "name": "BaseBdev3", 00:12:09.375 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:09.375 "is_configured": true, 00:12:09.375 "data_offset": 0, 00:12:09.375 "data_size": 65536 00:12:09.375 }, 00:12:09.375 { 00:12:09.375 "name": "BaseBdev4", 00:12:09.375 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:09.375 "is_configured": true, 00:12:09.375 "data_offset": 0, 00:12:09.375 "data_size": 65536 00:12:09.375 } 00:12:09.375 ] 00:12:09.375 }' 00:12:09.375 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.634 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:09.634 "name": "raid_bdev1", 00:12:09.634 "uuid": "16a9d4d9-5d90-4875-bb89-80a458594838", 00:12:09.634 "strip_size_kb": 0, 00:12:09.634 "state": "online", 00:12:09.634 "raid_level": "raid1", 00:12:09.634 "superblock": false, 00:12:09.634 "num_base_bdevs": 4, 00:12:09.634 "num_base_bdevs_discovered": 3, 00:12:09.634 "num_base_bdevs_operational": 3, 00:12:09.634 "base_bdevs_list": [ 00:12:09.634 { 00:12:09.634 "name": "spare", 00:12:09.634 "uuid": "49990394-c856-5bdb-997a-f3834441bc6e", 00:12:09.634 "is_configured": true, 00:12:09.634 "data_offset": 0, 00:12:09.634 "data_size": 65536 00:12:09.634 }, 00:12:09.634 { 00:12:09.634 "name": null, 00:12:09.634 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:09.634 "is_configured": false, 00:12:09.634 "data_offset": 0, 00:12:09.634 "data_size": 65536 00:12:09.634 }, 00:12:09.634 { 00:12:09.634 "name": "BaseBdev3", 00:12:09.635 "uuid": "15c66768-cbc8-587a-910b-1c55253f8ca9", 00:12:09.635 "is_configured": true, 00:12:09.635 "data_offset": 0, 00:12:09.635 "data_size": 65536 00:12:09.635 }, 00:12:09.635 { 00:12:09.635 "name": "BaseBdev4", 00:12:09.635 "uuid": "46f76ad5-f11e-5f3a-a355-db7cdde5d6b1", 00:12:09.635 "is_configured": true, 00:12:09.635 "data_offset": 0, 00:12:09.635 "data_size": 65536 00:12:09.635 } 00:12:09.635 ] 00:12:09.635 }' 00:12:09.635 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:09.635 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.894 [2024-09-29 16:50:31.512156] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:09.894 [2024-09-29 16:50:31.512227] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:09.894 [2024-09-29 16:50:31.512332] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:09.894 [2024-09-29 16:50:31.512462] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:09.894 [2024-09-29 16:50:31.512513] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:09.894 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:10.154 /dev/nbd0 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:10.154 1+0 records in 00:12:10.154 1+0 records out 00:12:10.154 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.0002267 s, 18.1 MB/s 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:10.154 16:50:31 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:10.414 /dev/nbd1 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:10.414 1+0 records in 00:12:10.414 1+0 records out 00:12:10.414 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000375582 s, 10.9 MB/s 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:10.414 16:50:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:10.674 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:10.675 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 87831 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 87831 ']' 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 87831 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:10.934 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87831 00:12:11.194 killing process with pid 87831 00:12:11.194 Received shutdown signal, test time was about 60.000000 seconds 00:12:11.194 00:12:11.194 Latency(us) 00:12:11.194 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:11.194 =================================================================================================================== 00:12:11.194 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:11.194 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:11.194 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:11.194 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87831' 00:12:11.194 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 87831 00:12:11.194 [2024-09-29 16:50:32.613313] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:11.194 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 87831 00:12:11.194 [2024-09-29 16:50:32.663253] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:11.454 00:12:11.454 real 0m15.034s 00:12:11.454 user 0m17.312s 00:12:11.454 sys 0m2.838s 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:11.454 ************************************ 00:12:11.454 END TEST raid_rebuild_test 00:12:11.454 ************************************ 00:12:11.454 16:50:32 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:12:11.454 16:50:32 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:11.454 16:50:32 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:11.454 16:50:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:11.454 ************************************ 00:12:11.454 START TEST raid_rebuild_test_sb 00:12:11.454 ************************************ 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true false true 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:11.454 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=88255 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 88255 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 88255 ']' 00:12:11.455 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:11.455 16:50:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:11.455 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:11.455 Zero copy mechanism will not be used. 00:12:11.455 [2024-09-29 16:50:33.078656] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:12:11.455 [2024-09-29 16:50:33.078801] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88255 ] 00:12:11.715 [2024-09-29 16:50:33.224861] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:11.715 [2024-09-29 16:50:33.270591] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:11.715 [2024-09-29 16:50:33.313636] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:11.715 [2024-09-29 16:50:33.313675] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.285 BaseBdev1_malloc 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.285 [2024-09-29 16:50:33.944190] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:12.285 [2024-09-29 16:50:33.944250] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.285 [2024-09-29 16:50:33.944282] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:12.285 [2024-09-29 16:50:33.944303] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.285 [2024-09-29 16:50:33.946401] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.285 [2024-09-29 16:50:33.946437] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:12.285 BaseBdev1 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.285 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.545 BaseBdev2_malloc 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.545 [2024-09-29 16:50:33.982897] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:12.545 [2024-09-29 16:50:33.982940] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.545 [2024-09-29 16:50:33.982982] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:12.545 [2024-09-29 16:50:33.982990] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.545 [2024-09-29 16:50:33.985070] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.545 [2024-09-29 16:50:33.985105] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:12.545 BaseBdev2 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.545 16:50:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.545 BaseBdev3_malloc 00:12:12.545 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.545 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:12.545 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 [2024-09-29 16:50:34.011574] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:12.546 [2024-09-29 16:50:34.011629] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.546 [2024-09-29 16:50:34.011654] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:12.546 [2024-09-29 16:50:34.011662] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.546 [2024-09-29 16:50:34.013686] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.546 [2024-09-29 16:50:34.013726] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:12.546 BaseBdev3 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 BaseBdev4_malloc 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 [2024-09-29 16:50:34.040392] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:12.546 [2024-09-29 16:50:34.040437] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.546 [2024-09-29 16:50:34.040458] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:12.546 [2024-09-29 16:50:34.040465] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.546 [2024-09-29 16:50:34.042494] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.546 [2024-09-29 16:50:34.042527] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:12.546 BaseBdev4 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 spare_malloc 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 spare_delay 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 [2024-09-29 16:50:34.080911] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:12.546 [2024-09-29 16:50:34.081006] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:12.546 [2024-09-29 16:50:34.081027] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:12.546 [2024-09-29 16:50:34.081036] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:12.546 [2024-09-29 16:50:34.083108] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:12.546 [2024-09-29 16:50:34.083142] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:12.546 spare 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 [2024-09-29 16:50:34.092967] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:12.546 [2024-09-29 16:50:34.094733] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:12.546 [2024-09-29 16:50:34.094810] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:12.546 [2024-09-29 16:50:34.094859] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:12.546 [2024-09-29 16:50:34.095023] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:12.546 [2024-09-29 16:50:34.095035] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:12.546 [2024-09-29 16:50:34.095312] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:12.546 [2024-09-29 16:50:34.095452] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:12.546 [2024-09-29 16:50:34.095465] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:12.546 [2024-09-29 16:50:34.095584] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:12.546 "name": "raid_bdev1", 00:12:12.546 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:12.546 "strip_size_kb": 0, 00:12:12.546 "state": "online", 00:12:12.546 "raid_level": "raid1", 00:12:12.546 "superblock": true, 00:12:12.546 "num_base_bdevs": 4, 00:12:12.546 "num_base_bdevs_discovered": 4, 00:12:12.546 "num_base_bdevs_operational": 4, 00:12:12.546 "base_bdevs_list": [ 00:12:12.546 { 00:12:12.546 "name": "BaseBdev1", 00:12:12.546 "uuid": "c976b3c9-8e19-594e-a72a-c855e7ec050f", 00:12:12.546 "is_configured": true, 00:12:12.546 "data_offset": 2048, 00:12:12.546 "data_size": 63488 00:12:12.546 }, 00:12:12.546 { 00:12:12.546 "name": "BaseBdev2", 00:12:12.546 "uuid": "1d646c1c-5fc4-5723-b551-c16c009224ed", 00:12:12.546 "is_configured": true, 00:12:12.546 "data_offset": 2048, 00:12:12.546 "data_size": 63488 00:12:12.546 }, 00:12:12.546 { 00:12:12.546 "name": "BaseBdev3", 00:12:12.546 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:12.546 "is_configured": true, 00:12:12.546 "data_offset": 2048, 00:12:12.546 "data_size": 63488 00:12:12.546 }, 00:12:12.546 { 00:12:12.546 "name": "BaseBdev4", 00:12:12.546 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:12.546 "is_configured": true, 00:12:12.546 "data_offset": 2048, 00:12:12.546 "data_size": 63488 00:12:12.546 } 00:12:12.546 ] 00:12:12.546 }' 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:12.546 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:13.116 [2024-09-29 16:50:34.540449] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:13.116 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:13.379 [2024-09-29 16:50:34.827681] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:13.379 /dev/nbd0 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:13.379 1+0 records in 00:12:13.379 1+0 records out 00:12:13.379 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000539459 s, 7.6 MB/s 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:13.379 16:50:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:12:18.657 63488+0 records in 00:12:18.657 63488+0 records out 00:12:18.657 32505856 bytes (33 MB, 31 MiB) copied, 5.39099 s, 6.0 MB/s 00:12:18.657 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:18.657 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:18.657 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:18.657 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:18.657 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:18.657 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:18.657 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:18.917 [2024-09-29 16:50:40.569570] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:18.917 16:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:18.917 [2024-09-29 16:50:40.585851] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:19.177 "name": "raid_bdev1", 00:12:19.177 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:19.177 "strip_size_kb": 0, 00:12:19.177 "state": "online", 00:12:19.177 "raid_level": "raid1", 00:12:19.177 "superblock": true, 00:12:19.177 "num_base_bdevs": 4, 00:12:19.177 "num_base_bdevs_discovered": 3, 00:12:19.177 "num_base_bdevs_operational": 3, 00:12:19.177 "base_bdevs_list": [ 00:12:19.177 { 00:12:19.177 "name": null, 00:12:19.177 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:19.177 "is_configured": false, 00:12:19.177 "data_offset": 0, 00:12:19.177 "data_size": 63488 00:12:19.177 }, 00:12:19.177 { 00:12:19.177 "name": "BaseBdev2", 00:12:19.177 "uuid": "1d646c1c-5fc4-5723-b551-c16c009224ed", 00:12:19.177 "is_configured": true, 00:12:19.177 "data_offset": 2048, 00:12:19.177 "data_size": 63488 00:12:19.177 }, 00:12:19.177 { 00:12:19.177 "name": "BaseBdev3", 00:12:19.177 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:19.177 "is_configured": true, 00:12:19.177 "data_offset": 2048, 00:12:19.177 "data_size": 63488 00:12:19.177 }, 00:12:19.177 { 00:12:19.177 "name": "BaseBdev4", 00:12:19.177 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:19.177 "is_configured": true, 00:12:19.177 "data_offset": 2048, 00:12:19.177 "data_size": 63488 00:12:19.177 } 00:12:19.177 ] 00:12:19.177 }' 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:19.177 16:50:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:19.436 16:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:19.437 16:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:19.437 16:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:19.437 [2024-09-29 16:50:41.065048] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:19.437 [2024-09-29 16:50:41.068466] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e420 00:12:19.437 [2024-09-29 16:50:41.070351] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:19.437 16:50:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:19.437 16:50:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:20.817 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:20.817 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:20.817 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:20.817 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:20.817 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:20.817 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:20.818 "name": "raid_bdev1", 00:12:20.818 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:20.818 "strip_size_kb": 0, 00:12:20.818 "state": "online", 00:12:20.818 "raid_level": "raid1", 00:12:20.818 "superblock": true, 00:12:20.818 "num_base_bdevs": 4, 00:12:20.818 "num_base_bdevs_discovered": 4, 00:12:20.818 "num_base_bdevs_operational": 4, 00:12:20.818 "process": { 00:12:20.818 "type": "rebuild", 00:12:20.818 "target": "spare", 00:12:20.818 "progress": { 00:12:20.818 "blocks": 20480, 00:12:20.818 "percent": 32 00:12:20.818 } 00:12:20.818 }, 00:12:20.818 "base_bdevs_list": [ 00:12:20.818 { 00:12:20.818 "name": "spare", 00:12:20.818 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:20.818 "is_configured": true, 00:12:20.818 "data_offset": 2048, 00:12:20.818 "data_size": 63488 00:12:20.818 }, 00:12:20.818 { 00:12:20.818 "name": "BaseBdev2", 00:12:20.818 "uuid": "1d646c1c-5fc4-5723-b551-c16c009224ed", 00:12:20.818 "is_configured": true, 00:12:20.818 "data_offset": 2048, 00:12:20.818 "data_size": 63488 00:12:20.818 }, 00:12:20.818 { 00:12:20.818 "name": "BaseBdev3", 00:12:20.818 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:20.818 "is_configured": true, 00:12:20.818 "data_offset": 2048, 00:12:20.818 "data_size": 63488 00:12:20.818 }, 00:12:20.818 { 00:12:20.818 "name": "BaseBdev4", 00:12:20.818 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:20.818 "is_configured": true, 00:12:20.818 "data_offset": 2048, 00:12:20.818 "data_size": 63488 00:12:20.818 } 00:12:20.818 ] 00:12:20.818 }' 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:20.818 [2024-09-29 16:50:42.213060] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:20.818 [2024-09-29 16:50:42.274954] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:20.818 [2024-09-29 16:50:42.275035] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:20.818 [2024-09-29 16:50:42.275054] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:20.818 [2024-09-29 16:50:42.275061] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:20.818 "name": "raid_bdev1", 00:12:20.818 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:20.818 "strip_size_kb": 0, 00:12:20.818 "state": "online", 00:12:20.818 "raid_level": "raid1", 00:12:20.818 "superblock": true, 00:12:20.818 "num_base_bdevs": 4, 00:12:20.818 "num_base_bdevs_discovered": 3, 00:12:20.818 "num_base_bdevs_operational": 3, 00:12:20.818 "base_bdevs_list": [ 00:12:20.818 { 00:12:20.818 "name": null, 00:12:20.818 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:20.818 "is_configured": false, 00:12:20.818 "data_offset": 0, 00:12:20.818 "data_size": 63488 00:12:20.818 }, 00:12:20.818 { 00:12:20.818 "name": "BaseBdev2", 00:12:20.818 "uuid": "1d646c1c-5fc4-5723-b551-c16c009224ed", 00:12:20.818 "is_configured": true, 00:12:20.818 "data_offset": 2048, 00:12:20.818 "data_size": 63488 00:12:20.818 }, 00:12:20.818 { 00:12:20.818 "name": "BaseBdev3", 00:12:20.818 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:20.818 "is_configured": true, 00:12:20.818 "data_offset": 2048, 00:12:20.818 "data_size": 63488 00:12:20.818 }, 00:12:20.818 { 00:12:20.818 "name": "BaseBdev4", 00:12:20.818 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:20.818 "is_configured": true, 00:12:20.818 "data_offset": 2048, 00:12:20.818 "data_size": 63488 00:12:20.818 } 00:12:20.818 ] 00:12:20.818 }' 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:20.818 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:21.078 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.337 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:21.338 "name": "raid_bdev1", 00:12:21.338 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:21.338 "strip_size_kb": 0, 00:12:21.338 "state": "online", 00:12:21.338 "raid_level": "raid1", 00:12:21.338 "superblock": true, 00:12:21.338 "num_base_bdevs": 4, 00:12:21.338 "num_base_bdevs_discovered": 3, 00:12:21.338 "num_base_bdevs_operational": 3, 00:12:21.338 "base_bdevs_list": [ 00:12:21.338 { 00:12:21.338 "name": null, 00:12:21.338 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:21.338 "is_configured": false, 00:12:21.338 "data_offset": 0, 00:12:21.338 "data_size": 63488 00:12:21.338 }, 00:12:21.338 { 00:12:21.338 "name": "BaseBdev2", 00:12:21.338 "uuid": "1d646c1c-5fc4-5723-b551-c16c009224ed", 00:12:21.338 "is_configured": true, 00:12:21.338 "data_offset": 2048, 00:12:21.338 "data_size": 63488 00:12:21.338 }, 00:12:21.338 { 00:12:21.338 "name": "BaseBdev3", 00:12:21.338 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:21.338 "is_configured": true, 00:12:21.338 "data_offset": 2048, 00:12:21.338 "data_size": 63488 00:12:21.338 }, 00:12:21.338 { 00:12:21.338 "name": "BaseBdev4", 00:12:21.338 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:21.338 "is_configured": true, 00:12:21.338 "data_offset": 2048, 00:12:21.338 "data_size": 63488 00:12:21.338 } 00:12:21.338 ] 00:12:21.338 }' 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:21.338 [2024-09-29 16:50:42.869965] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:21.338 [2024-09-29 16:50:42.873475] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e4f0 00:12:21.338 [2024-09-29 16:50:42.875472] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.338 16:50:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:22.315 "name": "raid_bdev1", 00:12:22.315 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:22.315 "strip_size_kb": 0, 00:12:22.315 "state": "online", 00:12:22.315 "raid_level": "raid1", 00:12:22.315 "superblock": true, 00:12:22.315 "num_base_bdevs": 4, 00:12:22.315 "num_base_bdevs_discovered": 4, 00:12:22.315 "num_base_bdevs_operational": 4, 00:12:22.315 "process": { 00:12:22.315 "type": "rebuild", 00:12:22.315 "target": "spare", 00:12:22.315 "progress": { 00:12:22.315 "blocks": 20480, 00:12:22.315 "percent": 32 00:12:22.315 } 00:12:22.315 }, 00:12:22.315 "base_bdevs_list": [ 00:12:22.315 { 00:12:22.315 "name": "spare", 00:12:22.315 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:22.315 "is_configured": true, 00:12:22.315 "data_offset": 2048, 00:12:22.315 "data_size": 63488 00:12:22.315 }, 00:12:22.315 { 00:12:22.315 "name": "BaseBdev2", 00:12:22.315 "uuid": "1d646c1c-5fc4-5723-b551-c16c009224ed", 00:12:22.315 "is_configured": true, 00:12:22.315 "data_offset": 2048, 00:12:22.315 "data_size": 63488 00:12:22.315 }, 00:12:22.315 { 00:12:22.315 "name": "BaseBdev3", 00:12:22.315 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:22.315 "is_configured": true, 00:12:22.315 "data_offset": 2048, 00:12:22.315 "data_size": 63488 00:12:22.315 }, 00:12:22.315 { 00:12:22.315 "name": "BaseBdev4", 00:12:22.315 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:22.315 "is_configured": true, 00:12:22.315 "data_offset": 2048, 00:12:22.315 "data_size": 63488 00:12:22.315 } 00:12:22.315 ] 00:12:22.315 }' 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.315 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:22.576 16:50:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:22.576 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.576 [2024-09-29 16:50:44.030019] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:22.576 [2024-09-29 16:50:44.179805] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000c3e4f0 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:22.576 "name": "raid_bdev1", 00:12:22.576 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:22.576 "strip_size_kb": 0, 00:12:22.576 "state": "online", 00:12:22.576 "raid_level": "raid1", 00:12:22.576 "superblock": true, 00:12:22.576 "num_base_bdevs": 4, 00:12:22.576 "num_base_bdevs_discovered": 3, 00:12:22.576 "num_base_bdevs_operational": 3, 00:12:22.576 "process": { 00:12:22.576 "type": "rebuild", 00:12:22.576 "target": "spare", 00:12:22.576 "progress": { 00:12:22.576 "blocks": 24576, 00:12:22.576 "percent": 38 00:12:22.576 } 00:12:22.576 }, 00:12:22.576 "base_bdevs_list": [ 00:12:22.576 { 00:12:22.576 "name": "spare", 00:12:22.576 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:22.576 "is_configured": true, 00:12:22.576 "data_offset": 2048, 00:12:22.576 "data_size": 63488 00:12:22.576 }, 00:12:22.576 { 00:12:22.576 "name": null, 00:12:22.576 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:22.576 "is_configured": false, 00:12:22.576 "data_offset": 0, 00:12:22.576 "data_size": 63488 00:12:22.576 }, 00:12:22.576 { 00:12:22.576 "name": "BaseBdev3", 00:12:22.576 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:22.576 "is_configured": true, 00:12:22.576 "data_offset": 2048, 00:12:22.576 "data_size": 63488 00:12:22.576 }, 00:12:22.576 { 00:12:22.576 "name": "BaseBdev4", 00:12:22.576 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:22.576 "is_configured": true, 00:12:22.576 "data_offset": 2048, 00:12:22.576 "data_size": 63488 00:12:22.576 } 00:12:22.576 ] 00:12:22.576 }' 00:12:22.576 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=369 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.836 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:22.836 "name": "raid_bdev1", 00:12:22.836 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:22.836 "strip_size_kb": 0, 00:12:22.836 "state": "online", 00:12:22.836 "raid_level": "raid1", 00:12:22.836 "superblock": true, 00:12:22.836 "num_base_bdevs": 4, 00:12:22.836 "num_base_bdevs_discovered": 3, 00:12:22.836 "num_base_bdevs_operational": 3, 00:12:22.836 "process": { 00:12:22.836 "type": "rebuild", 00:12:22.836 "target": "spare", 00:12:22.836 "progress": { 00:12:22.836 "blocks": 26624, 00:12:22.836 "percent": 41 00:12:22.836 } 00:12:22.836 }, 00:12:22.836 "base_bdevs_list": [ 00:12:22.836 { 00:12:22.836 "name": "spare", 00:12:22.836 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:22.836 "is_configured": true, 00:12:22.836 "data_offset": 2048, 00:12:22.836 "data_size": 63488 00:12:22.836 }, 00:12:22.836 { 00:12:22.836 "name": null, 00:12:22.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:22.837 "is_configured": false, 00:12:22.837 "data_offset": 0, 00:12:22.837 "data_size": 63488 00:12:22.837 }, 00:12:22.837 { 00:12:22.837 "name": "BaseBdev3", 00:12:22.837 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:22.837 "is_configured": true, 00:12:22.837 "data_offset": 2048, 00:12:22.837 "data_size": 63488 00:12:22.837 }, 00:12:22.837 { 00:12:22.837 "name": "BaseBdev4", 00:12:22.837 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:22.837 "is_configured": true, 00:12:22.837 "data_offset": 2048, 00:12:22.837 "data_size": 63488 00:12:22.837 } 00:12:22.837 ] 00:12:22.837 }' 00:12:22.837 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:22.837 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:22.837 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:22.837 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:22.837 16:50:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.220 "name": "raid_bdev1", 00:12:24.220 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:24.220 "strip_size_kb": 0, 00:12:24.220 "state": "online", 00:12:24.220 "raid_level": "raid1", 00:12:24.220 "superblock": true, 00:12:24.220 "num_base_bdevs": 4, 00:12:24.220 "num_base_bdevs_discovered": 3, 00:12:24.220 "num_base_bdevs_operational": 3, 00:12:24.220 "process": { 00:12:24.220 "type": "rebuild", 00:12:24.220 "target": "spare", 00:12:24.220 "progress": { 00:12:24.220 "blocks": 51200, 00:12:24.220 "percent": 80 00:12:24.220 } 00:12:24.220 }, 00:12:24.220 "base_bdevs_list": [ 00:12:24.220 { 00:12:24.220 "name": "spare", 00:12:24.220 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:24.220 "is_configured": true, 00:12:24.220 "data_offset": 2048, 00:12:24.220 "data_size": 63488 00:12:24.220 }, 00:12:24.220 { 00:12:24.220 "name": null, 00:12:24.220 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:24.220 "is_configured": false, 00:12:24.220 "data_offset": 0, 00:12:24.220 "data_size": 63488 00:12:24.220 }, 00:12:24.220 { 00:12:24.220 "name": "BaseBdev3", 00:12:24.220 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:24.220 "is_configured": true, 00:12:24.220 "data_offset": 2048, 00:12:24.220 "data_size": 63488 00:12:24.220 }, 00:12:24.220 { 00:12:24.220 "name": "BaseBdev4", 00:12:24.220 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:24.220 "is_configured": true, 00:12:24.220 "data_offset": 2048, 00:12:24.220 "data_size": 63488 00:12:24.220 } 00:12:24.220 ] 00:12:24.220 }' 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:24.220 16:50:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:24.480 [2024-09-29 16:50:46.086511] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:24.480 [2024-09-29 16:50:46.086630] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:24.480 [2024-09-29 16:50:46.086808] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:25.050 "name": "raid_bdev1", 00:12:25.050 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:25.050 "strip_size_kb": 0, 00:12:25.050 "state": "online", 00:12:25.050 "raid_level": "raid1", 00:12:25.050 "superblock": true, 00:12:25.050 "num_base_bdevs": 4, 00:12:25.050 "num_base_bdevs_discovered": 3, 00:12:25.050 "num_base_bdevs_operational": 3, 00:12:25.050 "base_bdevs_list": [ 00:12:25.050 { 00:12:25.050 "name": "spare", 00:12:25.050 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:25.050 "is_configured": true, 00:12:25.050 "data_offset": 2048, 00:12:25.050 "data_size": 63488 00:12:25.050 }, 00:12:25.050 { 00:12:25.050 "name": null, 00:12:25.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:25.050 "is_configured": false, 00:12:25.050 "data_offset": 0, 00:12:25.050 "data_size": 63488 00:12:25.050 }, 00:12:25.050 { 00:12:25.050 "name": "BaseBdev3", 00:12:25.050 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:25.050 "is_configured": true, 00:12:25.050 "data_offset": 2048, 00:12:25.050 "data_size": 63488 00:12:25.050 }, 00:12:25.050 { 00:12:25.050 "name": "BaseBdev4", 00:12:25.050 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:25.050 "is_configured": true, 00:12:25.050 "data_offset": 2048, 00:12:25.050 "data_size": 63488 00:12:25.050 } 00:12:25.050 ] 00:12:25.050 }' 00:12:25.050 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.309 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:25.309 "name": "raid_bdev1", 00:12:25.309 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:25.309 "strip_size_kb": 0, 00:12:25.309 "state": "online", 00:12:25.309 "raid_level": "raid1", 00:12:25.309 "superblock": true, 00:12:25.309 "num_base_bdevs": 4, 00:12:25.309 "num_base_bdevs_discovered": 3, 00:12:25.309 "num_base_bdevs_operational": 3, 00:12:25.310 "base_bdevs_list": [ 00:12:25.310 { 00:12:25.310 "name": "spare", 00:12:25.310 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:25.310 "is_configured": true, 00:12:25.310 "data_offset": 2048, 00:12:25.310 "data_size": 63488 00:12:25.310 }, 00:12:25.310 { 00:12:25.310 "name": null, 00:12:25.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:25.310 "is_configured": false, 00:12:25.310 "data_offset": 0, 00:12:25.310 "data_size": 63488 00:12:25.310 }, 00:12:25.310 { 00:12:25.310 "name": "BaseBdev3", 00:12:25.310 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:25.310 "is_configured": true, 00:12:25.310 "data_offset": 2048, 00:12:25.310 "data_size": 63488 00:12:25.310 }, 00:12:25.310 { 00:12:25.310 "name": "BaseBdev4", 00:12:25.310 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:25.310 "is_configured": true, 00:12:25.310 "data_offset": 2048, 00:12:25.310 "data_size": 63488 00:12:25.310 } 00:12:25.310 ] 00:12:25.310 }' 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:25.310 "name": "raid_bdev1", 00:12:25.310 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:25.310 "strip_size_kb": 0, 00:12:25.310 "state": "online", 00:12:25.310 "raid_level": "raid1", 00:12:25.310 "superblock": true, 00:12:25.310 "num_base_bdevs": 4, 00:12:25.310 "num_base_bdevs_discovered": 3, 00:12:25.310 "num_base_bdevs_operational": 3, 00:12:25.310 "base_bdevs_list": [ 00:12:25.310 { 00:12:25.310 "name": "spare", 00:12:25.310 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:25.310 "is_configured": true, 00:12:25.310 "data_offset": 2048, 00:12:25.310 "data_size": 63488 00:12:25.310 }, 00:12:25.310 { 00:12:25.310 "name": null, 00:12:25.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:25.310 "is_configured": false, 00:12:25.310 "data_offset": 0, 00:12:25.310 "data_size": 63488 00:12:25.310 }, 00:12:25.310 { 00:12:25.310 "name": "BaseBdev3", 00:12:25.310 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:25.310 "is_configured": true, 00:12:25.310 "data_offset": 2048, 00:12:25.310 "data_size": 63488 00:12:25.310 }, 00:12:25.310 { 00:12:25.310 "name": "BaseBdev4", 00:12:25.310 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:25.310 "is_configured": true, 00:12:25.310 "data_offset": 2048, 00:12:25.310 "data_size": 63488 00:12:25.310 } 00:12:25.310 ] 00:12:25.310 }' 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:25.310 16:50:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.878 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:25.878 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.879 [2024-09-29 16:50:47.352368] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:25.879 [2024-09-29 16:50:47.352396] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:25.879 [2024-09-29 16:50:47.352471] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:25.879 [2024-09-29 16:50:47.352552] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:25.879 [2024-09-29 16:50:47.352573] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:25.879 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:26.138 /dev/nbd0 00:12:26.138 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:26.139 1+0 records in 00:12:26.139 1+0 records out 00:12:26.139 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000285643 s, 14.3 MB/s 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:26.139 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:26.398 /dev/nbd1 00:12:26.398 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:26.398 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:26.398 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:26.399 1+0 records in 00:12:26.399 1+0 records out 00:12:26.399 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00354471 s, 1.2 MB/s 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:26.399 16:50:47 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:26.658 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.917 [2024-09-29 16:50:48.416721] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:26.917 [2024-09-29 16:50:48.416808] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:26.917 [2024-09-29 16:50:48.416829] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:12:26.917 [2024-09-29 16:50:48.416841] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:26.917 [2024-09-29 16:50:48.418932] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:26.917 [2024-09-29 16:50:48.418973] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:26.917 [2024-09-29 16:50:48.419063] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:26.917 [2024-09-29 16:50:48.419102] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:26.917 [2024-09-29 16:50:48.419200] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:26.917 [2024-09-29 16:50:48.419295] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:26.917 spare 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.917 [2024-09-29 16:50:48.519179] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:26.917 [2024-09-29 16:50:48.519208] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:26.917 [2024-09-29 16:50:48.519477] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:12:26.917 [2024-09-29 16:50:48.519608] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:26.917 [2024-09-29 16:50:48.519620] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:26.917 [2024-09-29 16:50:48.519804] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:26.917 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:26.918 "name": "raid_bdev1", 00:12:26.918 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:26.918 "strip_size_kb": 0, 00:12:26.918 "state": "online", 00:12:26.918 "raid_level": "raid1", 00:12:26.918 "superblock": true, 00:12:26.918 "num_base_bdevs": 4, 00:12:26.918 "num_base_bdevs_discovered": 3, 00:12:26.918 "num_base_bdevs_operational": 3, 00:12:26.918 "base_bdevs_list": [ 00:12:26.918 { 00:12:26.918 "name": "spare", 00:12:26.918 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:26.918 "is_configured": true, 00:12:26.918 "data_offset": 2048, 00:12:26.918 "data_size": 63488 00:12:26.918 }, 00:12:26.918 { 00:12:26.918 "name": null, 00:12:26.918 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:26.918 "is_configured": false, 00:12:26.918 "data_offset": 2048, 00:12:26.918 "data_size": 63488 00:12:26.918 }, 00:12:26.918 { 00:12:26.918 "name": "BaseBdev3", 00:12:26.918 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:26.918 "is_configured": true, 00:12:26.918 "data_offset": 2048, 00:12:26.918 "data_size": 63488 00:12:26.918 }, 00:12:26.918 { 00:12:26.918 "name": "BaseBdev4", 00:12:26.918 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:26.918 "is_configured": true, 00:12:26.918 "data_offset": 2048, 00:12:26.918 "data_size": 63488 00:12:26.918 } 00:12:26.918 ] 00:12:26.918 }' 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:26.918 16:50:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:27.486 "name": "raid_bdev1", 00:12:27.486 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:27.486 "strip_size_kb": 0, 00:12:27.486 "state": "online", 00:12:27.486 "raid_level": "raid1", 00:12:27.486 "superblock": true, 00:12:27.486 "num_base_bdevs": 4, 00:12:27.486 "num_base_bdevs_discovered": 3, 00:12:27.486 "num_base_bdevs_operational": 3, 00:12:27.486 "base_bdevs_list": [ 00:12:27.486 { 00:12:27.486 "name": "spare", 00:12:27.486 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:27.486 "is_configured": true, 00:12:27.486 "data_offset": 2048, 00:12:27.486 "data_size": 63488 00:12:27.486 }, 00:12:27.486 { 00:12:27.486 "name": null, 00:12:27.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.486 "is_configured": false, 00:12:27.486 "data_offset": 2048, 00:12:27.486 "data_size": 63488 00:12:27.486 }, 00:12:27.486 { 00:12:27.486 "name": "BaseBdev3", 00:12:27.486 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:27.486 "is_configured": true, 00:12:27.486 "data_offset": 2048, 00:12:27.486 "data_size": 63488 00:12:27.486 }, 00:12:27.486 { 00:12:27.486 "name": "BaseBdev4", 00:12:27.486 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:27.486 "is_configured": true, 00:12:27.486 "data_offset": 2048, 00:12:27.486 "data_size": 63488 00:12:27.486 } 00:12:27.486 ] 00:12:27.486 }' 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:27.486 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.745 [2024-09-29 16:50:49.175460] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:27.745 "name": "raid_bdev1", 00:12:27.745 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:27.745 "strip_size_kb": 0, 00:12:27.745 "state": "online", 00:12:27.745 "raid_level": "raid1", 00:12:27.745 "superblock": true, 00:12:27.745 "num_base_bdevs": 4, 00:12:27.745 "num_base_bdevs_discovered": 2, 00:12:27.745 "num_base_bdevs_operational": 2, 00:12:27.745 "base_bdevs_list": [ 00:12:27.745 { 00:12:27.745 "name": null, 00:12:27.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.745 "is_configured": false, 00:12:27.745 "data_offset": 0, 00:12:27.745 "data_size": 63488 00:12:27.745 }, 00:12:27.745 { 00:12:27.745 "name": null, 00:12:27.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:27.745 "is_configured": false, 00:12:27.745 "data_offset": 2048, 00:12:27.745 "data_size": 63488 00:12:27.745 }, 00:12:27.745 { 00:12:27.745 "name": "BaseBdev3", 00:12:27.745 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:27.745 "is_configured": true, 00:12:27.745 "data_offset": 2048, 00:12:27.745 "data_size": 63488 00:12:27.745 }, 00:12:27.745 { 00:12:27.745 "name": "BaseBdev4", 00:12:27.745 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:27.745 "is_configured": true, 00:12:27.745 "data_offset": 2048, 00:12:27.745 "data_size": 63488 00:12:27.745 } 00:12:27.745 ] 00:12:27.745 }' 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:27.745 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.004 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:28.004 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.004 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:28.004 [2024-09-29 16:50:49.602868] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:28.004 [2024-09-29 16:50:49.603040] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:28.004 [2024-09-29 16:50:49.603060] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:28.004 [2024-09-29 16:50:49.603103] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:28.004 [2024-09-29 16:50:49.606370] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caebd0 00:12:28.004 [2024-09-29 16:50:49.608371] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:28.004 16:50:49 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.004 16:50:49 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:29.383 "name": "raid_bdev1", 00:12:29.383 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:29.383 "strip_size_kb": 0, 00:12:29.383 "state": "online", 00:12:29.383 "raid_level": "raid1", 00:12:29.383 "superblock": true, 00:12:29.383 "num_base_bdevs": 4, 00:12:29.383 "num_base_bdevs_discovered": 3, 00:12:29.383 "num_base_bdevs_operational": 3, 00:12:29.383 "process": { 00:12:29.383 "type": "rebuild", 00:12:29.383 "target": "spare", 00:12:29.383 "progress": { 00:12:29.383 "blocks": 20480, 00:12:29.383 "percent": 32 00:12:29.383 } 00:12:29.383 }, 00:12:29.383 "base_bdevs_list": [ 00:12:29.383 { 00:12:29.383 "name": "spare", 00:12:29.383 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:29.383 "is_configured": true, 00:12:29.383 "data_offset": 2048, 00:12:29.383 "data_size": 63488 00:12:29.383 }, 00:12:29.383 { 00:12:29.383 "name": null, 00:12:29.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.383 "is_configured": false, 00:12:29.383 "data_offset": 2048, 00:12:29.383 "data_size": 63488 00:12:29.383 }, 00:12:29.383 { 00:12:29.383 "name": "BaseBdev3", 00:12:29.383 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:29.383 "is_configured": true, 00:12:29.383 "data_offset": 2048, 00:12:29.383 "data_size": 63488 00:12:29.383 }, 00:12:29.383 { 00:12:29.383 "name": "BaseBdev4", 00:12:29.383 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:29.383 "is_configured": true, 00:12:29.383 "data_offset": 2048, 00:12:29.383 "data_size": 63488 00:12:29.383 } 00:12:29.383 ] 00:12:29.383 }' 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.383 [2024-09-29 16:50:50.775185] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.383 [2024-09-29 16:50:50.812367] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:29.383 [2024-09-29 16:50:50.812499] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:29.383 [2024-09-29 16:50:50.812534] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:29.383 [2024-09-29 16:50:50.812557] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:29.383 "name": "raid_bdev1", 00:12:29.383 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:29.383 "strip_size_kb": 0, 00:12:29.383 "state": "online", 00:12:29.383 "raid_level": "raid1", 00:12:29.383 "superblock": true, 00:12:29.383 "num_base_bdevs": 4, 00:12:29.383 "num_base_bdevs_discovered": 2, 00:12:29.383 "num_base_bdevs_operational": 2, 00:12:29.383 "base_bdevs_list": [ 00:12:29.383 { 00:12:29.383 "name": null, 00:12:29.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.383 "is_configured": false, 00:12:29.383 "data_offset": 0, 00:12:29.383 "data_size": 63488 00:12:29.383 }, 00:12:29.383 { 00:12:29.383 "name": null, 00:12:29.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:29.383 "is_configured": false, 00:12:29.383 "data_offset": 2048, 00:12:29.383 "data_size": 63488 00:12:29.383 }, 00:12:29.383 { 00:12:29.383 "name": "BaseBdev3", 00:12:29.383 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:29.383 "is_configured": true, 00:12:29.383 "data_offset": 2048, 00:12:29.383 "data_size": 63488 00:12:29.383 }, 00:12:29.383 { 00:12:29.383 "name": "BaseBdev4", 00:12:29.383 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:29.383 "is_configured": true, 00:12:29.383 "data_offset": 2048, 00:12:29.383 "data_size": 63488 00:12:29.383 } 00:12:29.383 ] 00:12:29.383 }' 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:29.383 16:50:50 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.642 16:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:29.642 16:50:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.642 16:50:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:29.642 [2024-09-29 16:50:51.275441] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:29.642 [2024-09-29 16:50:51.275542] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:29.642 [2024-09-29 16:50:51.275569] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:12:29.642 [2024-09-29 16:50:51.275581] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:29.642 [2024-09-29 16:50:51.276018] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:29.642 [2024-09-29 16:50:51.276039] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:29.642 [2024-09-29 16:50:51.276116] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:29.642 [2024-09-29 16:50:51.276132] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:29.642 [2024-09-29 16:50:51.276141] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:29.642 [2024-09-29 16:50:51.276165] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:29.642 spare 00:12:29.642 [2024-09-29 16:50:51.279190] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeca0 00:12:29.643 [2024-09-29 16:50:51.281046] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:29.643 16:50:51 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.643 16:50:51 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.022 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.022 "name": "raid_bdev1", 00:12:31.022 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:31.022 "strip_size_kb": 0, 00:12:31.022 "state": "online", 00:12:31.022 "raid_level": "raid1", 00:12:31.022 "superblock": true, 00:12:31.022 "num_base_bdevs": 4, 00:12:31.022 "num_base_bdevs_discovered": 3, 00:12:31.022 "num_base_bdevs_operational": 3, 00:12:31.022 "process": { 00:12:31.022 "type": "rebuild", 00:12:31.022 "target": "spare", 00:12:31.022 "progress": { 00:12:31.022 "blocks": 20480, 00:12:31.022 "percent": 32 00:12:31.022 } 00:12:31.022 }, 00:12:31.022 "base_bdevs_list": [ 00:12:31.022 { 00:12:31.022 "name": "spare", 00:12:31.022 "uuid": "400f9371-4fcf-5694-bb91-fea5cfd6e8b7", 00:12:31.022 "is_configured": true, 00:12:31.022 "data_offset": 2048, 00:12:31.022 "data_size": 63488 00:12:31.022 }, 00:12:31.022 { 00:12:31.022 "name": null, 00:12:31.022 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.022 "is_configured": false, 00:12:31.022 "data_offset": 2048, 00:12:31.022 "data_size": 63488 00:12:31.022 }, 00:12:31.022 { 00:12:31.022 "name": "BaseBdev3", 00:12:31.022 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:31.022 "is_configured": true, 00:12:31.022 "data_offset": 2048, 00:12:31.022 "data_size": 63488 00:12:31.022 }, 00:12:31.022 { 00:12:31.022 "name": "BaseBdev4", 00:12:31.022 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:31.022 "is_configured": true, 00:12:31.022 "data_offset": 2048, 00:12:31.023 "data_size": 63488 00:12:31.023 } 00:12:31.023 ] 00:12:31.023 }' 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.023 [2024-09-29 16:50:52.440244] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:31.023 [2024-09-29 16:50:52.485048] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:31.023 [2024-09-29 16:50:52.485116] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:31.023 [2024-09-29 16:50:52.485134] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:31.023 [2024-09-29 16:50:52.485141] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:31.023 "name": "raid_bdev1", 00:12:31.023 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:31.023 "strip_size_kb": 0, 00:12:31.023 "state": "online", 00:12:31.023 "raid_level": "raid1", 00:12:31.023 "superblock": true, 00:12:31.023 "num_base_bdevs": 4, 00:12:31.023 "num_base_bdevs_discovered": 2, 00:12:31.023 "num_base_bdevs_operational": 2, 00:12:31.023 "base_bdevs_list": [ 00:12:31.023 { 00:12:31.023 "name": null, 00:12:31.023 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.023 "is_configured": false, 00:12:31.023 "data_offset": 0, 00:12:31.023 "data_size": 63488 00:12:31.023 }, 00:12:31.023 { 00:12:31.023 "name": null, 00:12:31.023 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.023 "is_configured": false, 00:12:31.023 "data_offset": 2048, 00:12:31.023 "data_size": 63488 00:12:31.023 }, 00:12:31.023 { 00:12:31.023 "name": "BaseBdev3", 00:12:31.023 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:31.023 "is_configured": true, 00:12:31.023 "data_offset": 2048, 00:12:31.023 "data_size": 63488 00:12:31.023 }, 00:12:31.023 { 00:12:31.023 "name": "BaseBdev4", 00:12:31.023 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:31.023 "is_configured": true, 00:12:31.023 "data_offset": 2048, 00:12:31.023 "data_size": 63488 00:12:31.023 } 00:12:31.023 ] 00:12:31.023 }' 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:31.023 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:31.283 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.542 16:50:52 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.542 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:31.542 "name": "raid_bdev1", 00:12:31.542 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:31.542 "strip_size_kb": 0, 00:12:31.542 "state": "online", 00:12:31.542 "raid_level": "raid1", 00:12:31.542 "superblock": true, 00:12:31.542 "num_base_bdevs": 4, 00:12:31.542 "num_base_bdevs_discovered": 2, 00:12:31.542 "num_base_bdevs_operational": 2, 00:12:31.542 "base_bdevs_list": [ 00:12:31.542 { 00:12:31.542 "name": null, 00:12:31.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.542 "is_configured": false, 00:12:31.542 "data_offset": 0, 00:12:31.542 "data_size": 63488 00:12:31.542 }, 00:12:31.542 { 00:12:31.542 "name": null, 00:12:31.542 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.542 "is_configured": false, 00:12:31.542 "data_offset": 2048, 00:12:31.542 "data_size": 63488 00:12:31.542 }, 00:12:31.542 { 00:12:31.542 "name": "BaseBdev3", 00:12:31.542 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:31.542 "is_configured": true, 00:12:31.542 "data_offset": 2048, 00:12:31.542 "data_size": 63488 00:12:31.542 }, 00:12:31.542 { 00:12:31.542 "name": "BaseBdev4", 00:12:31.542 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:31.542 "is_configured": true, 00:12:31.542 "data_offset": 2048, 00:12:31.542 "data_size": 63488 00:12:31.542 } 00:12:31.542 ] 00:12:31.542 }' 00:12:31.542 16:50:52 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:31.542 [2024-09-29 16:50:53.095612] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:31.542 [2024-09-29 16:50:53.095709] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:31.542 [2024-09-29 16:50:53.095759] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:12:31.542 [2024-09-29 16:50:53.095790] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:31.542 [2024-09-29 16:50:53.096201] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:31.542 [2024-09-29 16:50:53.096257] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:31.542 [2024-09-29 16:50:53.096368] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:31.542 [2024-09-29 16:50:53.096409] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:31.542 [2024-09-29 16:50:53.096479] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:31.542 [2024-09-29 16:50:53.096512] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:31.542 BaseBdev1 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.542 16:50:53 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:32.482 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:32.483 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:32.483 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:32.483 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.483 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:32.483 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:32.483 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:32.483 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:32.743 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:32.743 "name": "raid_bdev1", 00:12:32.743 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:32.743 "strip_size_kb": 0, 00:12:32.743 "state": "online", 00:12:32.743 "raid_level": "raid1", 00:12:32.743 "superblock": true, 00:12:32.743 "num_base_bdevs": 4, 00:12:32.743 "num_base_bdevs_discovered": 2, 00:12:32.743 "num_base_bdevs_operational": 2, 00:12:32.743 "base_bdevs_list": [ 00:12:32.743 { 00:12:32.743 "name": null, 00:12:32.743 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:32.743 "is_configured": false, 00:12:32.743 "data_offset": 0, 00:12:32.743 "data_size": 63488 00:12:32.743 }, 00:12:32.743 { 00:12:32.743 "name": null, 00:12:32.743 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:32.743 "is_configured": false, 00:12:32.743 "data_offset": 2048, 00:12:32.743 "data_size": 63488 00:12:32.743 }, 00:12:32.743 { 00:12:32.743 "name": "BaseBdev3", 00:12:32.743 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:32.743 "is_configured": true, 00:12:32.743 "data_offset": 2048, 00:12:32.743 "data_size": 63488 00:12:32.743 }, 00:12:32.743 { 00:12:32.743 "name": "BaseBdev4", 00:12:32.743 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:32.743 "is_configured": true, 00:12:32.743 "data_offset": 2048, 00:12:32.743 "data_size": 63488 00:12:32.743 } 00:12:32.743 ] 00:12:32.743 }' 00:12:32.743 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:32.743 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:33.002 "name": "raid_bdev1", 00:12:33.002 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:33.002 "strip_size_kb": 0, 00:12:33.002 "state": "online", 00:12:33.002 "raid_level": "raid1", 00:12:33.002 "superblock": true, 00:12:33.002 "num_base_bdevs": 4, 00:12:33.002 "num_base_bdevs_discovered": 2, 00:12:33.002 "num_base_bdevs_operational": 2, 00:12:33.002 "base_bdevs_list": [ 00:12:33.002 { 00:12:33.002 "name": null, 00:12:33.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.002 "is_configured": false, 00:12:33.002 "data_offset": 0, 00:12:33.002 "data_size": 63488 00:12:33.002 }, 00:12:33.002 { 00:12:33.002 "name": null, 00:12:33.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:33.002 "is_configured": false, 00:12:33.002 "data_offset": 2048, 00:12:33.002 "data_size": 63488 00:12:33.002 }, 00:12:33.002 { 00:12:33.002 "name": "BaseBdev3", 00:12:33.002 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:33.002 "is_configured": true, 00:12:33.002 "data_offset": 2048, 00:12:33.002 "data_size": 63488 00:12:33.002 }, 00:12:33.002 { 00:12:33.002 "name": "BaseBdev4", 00:12:33.002 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:33.002 "is_configured": true, 00:12:33.002 "data_offset": 2048, 00:12:33.002 "data_size": 63488 00:12:33.002 } 00:12:33.002 ] 00:12:33.002 }' 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:33.002 [2024-09-29 16:50:54.661044] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:33.002 [2024-09-29 16:50:54.661249] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:12:33.002 [2024-09-29 16:50:54.661316] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:33.002 request: 00:12:33.002 { 00:12:33.002 "base_bdev": "BaseBdev1", 00:12:33.002 "raid_bdev": "raid_bdev1", 00:12:33.002 "method": "bdev_raid_add_base_bdev", 00:12:33.002 "req_id": 1 00:12:33.002 } 00:12:33.002 Got JSON-RPC error response 00:12:33.002 response: 00:12:33.002 { 00:12:33.002 "code": -22, 00:12:33.002 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:33.002 } 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:33.002 16:50:54 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.382 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:34.382 "name": "raid_bdev1", 00:12:34.382 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:34.382 "strip_size_kb": 0, 00:12:34.382 "state": "online", 00:12:34.382 "raid_level": "raid1", 00:12:34.382 "superblock": true, 00:12:34.382 "num_base_bdevs": 4, 00:12:34.382 "num_base_bdevs_discovered": 2, 00:12:34.382 "num_base_bdevs_operational": 2, 00:12:34.382 "base_bdevs_list": [ 00:12:34.382 { 00:12:34.382 "name": null, 00:12:34.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.383 "is_configured": false, 00:12:34.383 "data_offset": 0, 00:12:34.383 "data_size": 63488 00:12:34.383 }, 00:12:34.383 { 00:12:34.383 "name": null, 00:12:34.383 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.383 "is_configured": false, 00:12:34.383 "data_offset": 2048, 00:12:34.383 "data_size": 63488 00:12:34.383 }, 00:12:34.383 { 00:12:34.383 "name": "BaseBdev3", 00:12:34.383 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:34.383 "is_configured": true, 00:12:34.383 "data_offset": 2048, 00:12:34.383 "data_size": 63488 00:12:34.383 }, 00:12:34.383 { 00:12:34.383 "name": "BaseBdev4", 00:12:34.383 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:34.383 "is_configured": true, 00:12:34.383 "data_offset": 2048, 00:12:34.383 "data_size": 63488 00:12:34.383 } 00:12:34.383 ] 00:12:34.383 }' 00:12:34.383 16:50:55 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:34.383 16:50:55 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.643 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:34.643 "name": "raid_bdev1", 00:12:34.643 "uuid": "d88c83ee-3c69-40f4-b488-98797921d81d", 00:12:34.643 "strip_size_kb": 0, 00:12:34.643 "state": "online", 00:12:34.643 "raid_level": "raid1", 00:12:34.643 "superblock": true, 00:12:34.643 "num_base_bdevs": 4, 00:12:34.643 "num_base_bdevs_discovered": 2, 00:12:34.643 "num_base_bdevs_operational": 2, 00:12:34.643 "base_bdevs_list": [ 00:12:34.643 { 00:12:34.643 "name": null, 00:12:34.643 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.643 "is_configured": false, 00:12:34.643 "data_offset": 0, 00:12:34.643 "data_size": 63488 00:12:34.643 }, 00:12:34.643 { 00:12:34.643 "name": null, 00:12:34.644 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.644 "is_configured": false, 00:12:34.644 "data_offset": 2048, 00:12:34.644 "data_size": 63488 00:12:34.644 }, 00:12:34.644 { 00:12:34.644 "name": "BaseBdev3", 00:12:34.644 "uuid": "4d3dd1a6-02f1-5f2e-a413-2d198c7eb652", 00:12:34.644 "is_configured": true, 00:12:34.644 "data_offset": 2048, 00:12:34.644 "data_size": 63488 00:12:34.644 }, 00:12:34.644 { 00:12:34.644 "name": "BaseBdev4", 00:12:34.644 "uuid": "064b8732-11d1-5f93-ba13-049c53b495ad", 00:12:34.644 "is_configured": true, 00:12:34.644 "data_offset": 2048, 00:12:34.644 "data_size": 63488 00:12:34.644 } 00:12:34.644 ] 00:12:34.644 }' 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 88255 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 88255 ']' 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 88255 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88255 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88255' 00:12:34.644 killing process with pid 88255 00:12:34.644 Received shutdown signal, test time was about 60.000000 seconds 00:12:34.644 00:12:34.644 Latency(us) 00:12:34.644 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:34.644 =================================================================================================================== 00:12:34.644 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 88255 00:12:34.644 [2024-09-29 16:50:56.259055] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:34.644 [2024-09-29 16:50:56.259184] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:34.644 [2024-09-29 16:50:56.259247] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:34.644 [2024-09-29 16:50:56.259259] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:34.644 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 88255 00:12:34.644 [2024-09-29 16:50:56.309302] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:34.905 16:50:56 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:34.905 00:12:34.905 real 0m23.562s 00:12:34.905 user 0m28.592s 00:12:34.905 sys 0m4.140s 00:12:34.905 ************************************ 00:12:34.905 END TEST raid_rebuild_test_sb 00:12:34.905 ************************************ 00:12:34.905 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:34.905 16:50:56 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:35.164 16:50:56 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:12:35.164 16:50:56 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:35.164 16:50:56 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:35.164 16:50:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:35.164 ************************************ 00:12:35.164 START TEST raid_rebuild_test_io 00:12:35.164 ************************************ 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false true true 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=88991 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 88991 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 88991 ']' 00:12:35.164 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:35.164 16:50:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.164 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:35.164 Zero copy mechanism will not be used. 00:12:35.164 [2024-09-29 16:50:56.725515] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:12:35.164 [2024-09-29 16:50:56.726109] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88991 ] 00:12:35.424 [2024-09-29 16:50:56.871368] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:35.424 [2024-09-29 16:50:56.920180] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:35.424 [2024-09-29 16:50:56.964089] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:35.424 [2024-09-29 16:50:56.964122] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 BaseBdev1_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 [2024-09-29 16:50:57.558581] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:35.995 [2024-09-29 16:50:57.558699] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:35.995 [2024-09-29 16:50:57.558756] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:35.995 [2024-09-29 16:50:57.558831] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:35.995 [2024-09-29 16:50:57.560965] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:35.995 [2024-09-29 16:50:57.561035] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:35.995 BaseBdev1 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 BaseBdev2_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 [2024-09-29 16:50:57.600811] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:35.995 [2024-09-29 16:50:57.600905] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:35.995 [2024-09-29 16:50:57.600948] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:35.995 [2024-09-29 16:50:57.600967] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:35.995 [2024-09-29 16:50:57.605504] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:35.995 [2024-09-29 16:50:57.605574] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:35.995 BaseBdev2 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 BaseBdev3_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 [2024-09-29 16:50:57.631690] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:35.995 [2024-09-29 16:50:57.631821] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:35.995 [2024-09-29 16:50:57.631869] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:35.995 [2024-09-29 16:50:57.631906] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:35.995 [2024-09-29 16:50:57.634026] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:35.995 [2024-09-29 16:50:57.634094] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:35.995 BaseBdev3 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 BaseBdev4_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.995 [2024-09-29 16:50:57.660575] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:35.995 [2024-09-29 16:50:57.660675] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:35.995 [2024-09-29 16:50:57.660712] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:35.995 [2024-09-29 16:50:57.660752] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:35.995 [2024-09-29 16:50:57.662858] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:35.995 [2024-09-29 16:50:57.662924] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:35.995 BaseBdev4 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.995 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.254 spare_malloc 00:12:36.254 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.255 spare_delay 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.255 [2024-09-29 16:50:57.701186] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:36.255 [2024-09-29 16:50:57.701282] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:36.255 [2024-09-29 16:50:57.701318] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:36.255 [2024-09-29 16:50:57.701345] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:36.255 [2024-09-29 16:50:57.703471] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:36.255 [2024-09-29 16:50:57.703541] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:36.255 spare 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.255 [2024-09-29 16:50:57.713233] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:36.255 [2024-09-29 16:50:57.715050] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:36.255 [2024-09-29 16:50:57.715112] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:36.255 [2024-09-29 16:50:57.715157] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:36.255 [2024-09-29 16:50:57.715230] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:36.255 [2024-09-29 16:50:57.715238] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:36.255 [2024-09-29 16:50:57.715477] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:36.255 [2024-09-29 16:50:57.715596] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:36.255 [2024-09-29 16:50:57.715609] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:36.255 [2024-09-29 16:50:57.715719] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.255 "name": "raid_bdev1", 00:12:36.255 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:36.255 "strip_size_kb": 0, 00:12:36.255 "state": "online", 00:12:36.255 "raid_level": "raid1", 00:12:36.255 "superblock": false, 00:12:36.255 "num_base_bdevs": 4, 00:12:36.255 "num_base_bdevs_discovered": 4, 00:12:36.255 "num_base_bdevs_operational": 4, 00:12:36.255 "base_bdevs_list": [ 00:12:36.255 { 00:12:36.255 "name": "BaseBdev1", 00:12:36.255 "uuid": "3c1a20df-6643-50eb-b755-a0a65b5b27f8", 00:12:36.255 "is_configured": true, 00:12:36.255 "data_offset": 0, 00:12:36.255 "data_size": 65536 00:12:36.255 }, 00:12:36.255 { 00:12:36.255 "name": "BaseBdev2", 00:12:36.255 "uuid": "7fedfea9-fe9c-5c07-adab-f829c7027c43", 00:12:36.255 "is_configured": true, 00:12:36.255 "data_offset": 0, 00:12:36.255 "data_size": 65536 00:12:36.255 }, 00:12:36.255 { 00:12:36.255 "name": "BaseBdev3", 00:12:36.255 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:36.255 "is_configured": true, 00:12:36.255 "data_offset": 0, 00:12:36.255 "data_size": 65536 00:12:36.255 }, 00:12:36.255 { 00:12:36.255 "name": "BaseBdev4", 00:12:36.255 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:36.255 "is_configured": true, 00:12:36.255 "data_offset": 0, 00:12:36.255 "data_size": 65536 00:12:36.255 } 00:12:36.255 ] 00:12:36.255 }' 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.255 16:50:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.825 [2024-09-29 16:50:58.196703] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:36.825 [2024-09-29 16:50:58.292187] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:36.825 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:36.826 "name": "raid_bdev1", 00:12:36.826 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:36.826 "strip_size_kb": 0, 00:12:36.826 "state": "online", 00:12:36.826 "raid_level": "raid1", 00:12:36.826 "superblock": false, 00:12:36.826 "num_base_bdevs": 4, 00:12:36.826 "num_base_bdevs_discovered": 3, 00:12:36.826 "num_base_bdevs_operational": 3, 00:12:36.826 "base_bdevs_list": [ 00:12:36.826 { 00:12:36.826 "name": null, 00:12:36.826 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.826 "is_configured": false, 00:12:36.826 "data_offset": 0, 00:12:36.826 "data_size": 65536 00:12:36.826 }, 00:12:36.826 { 00:12:36.826 "name": "BaseBdev2", 00:12:36.826 "uuid": "7fedfea9-fe9c-5c07-adab-f829c7027c43", 00:12:36.826 "is_configured": true, 00:12:36.826 "data_offset": 0, 00:12:36.826 "data_size": 65536 00:12:36.826 }, 00:12:36.826 { 00:12:36.826 "name": "BaseBdev3", 00:12:36.826 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:36.826 "is_configured": true, 00:12:36.826 "data_offset": 0, 00:12:36.826 "data_size": 65536 00:12:36.826 }, 00:12:36.826 { 00:12:36.826 "name": "BaseBdev4", 00:12:36.826 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:36.826 "is_configured": true, 00:12:36.826 "data_offset": 0, 00:12:36.826 "data_size": 65536 00:12:36.826 } 00:12:36.826 ] 00:12:36.826 }' 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:36.826 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.826 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:36.826 Zero copy mechanism will not be used. 00:12:36.826 Running I/O for 60 seconds... 00:12:36.826 [2024-09-29 16:50:58.381995] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:37.086 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:37.086 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:37.086 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:37.086 [2024-09-29 16:50:58.727350] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:37.346 16:50:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:37.346 16:50:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:37.346 [2024-09-29 16:50:58.770160] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:12:37.346 [2024-09-29 16:50:58.772208] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:37.346 [2024-09-29 16:50:58.920302] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:37.605 [2024-09-29 16:50:59.170817] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:37.865 209.00 IOPS, 627.00 MiB/s [2024-09-29 16:50:59.441005] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:37.865 [2024-09-29 16:50:59.442018] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:38.124 [2024-09-29 16:50:59.665859] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:38.125 [2024-09-29 16:50:59.666114] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.125 16:50:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.407 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:38.407 "name": "raid_bdev1", 00:12:38.407 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:38.408 "strip_size_kb": 0, 00:12:38.408 "state": "online", 00:12:38.408 "raid_level": "raid1", 00:12:38.408 "superblock": false, 00:12:38.408 "num_base_bdevs": 4, 00:12:38.408 "num_base_bdevs_discovered": 4, 00:12:38.408 "num_base_bdevs_operational": 4, 00:12:38.408 "process": { 00:12:38.408 "type": "rebuild", 00:12:38.408 "target": "spare", 00:12:38.408 "progress": { 00:12:38.408 "blocks": 10240, 00:12:38.408 "percent": 15 00:12:38.408 } 00:12:38.408 }, 00:12:38.408 "base_bdevs_list": [ 00:12:38.408 { 00:12:38.408 "name": "spare", 00:12:38.408 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:38.408 "is_configured": true, 00:12:38.408 "data_offset": 0, 00:12:38.408 "data_size": 65536 00:12:38.408 }, 00:12:38.408 { 00:12:38.408 "name": "BaseBdev2", 00:12:38.408 "uuid": "7fedfea9-fe9c-5c07-adab-f829c7027c43", 00:12:38.408 "is_configured": true, 00:12:38.408 "data_offset": 0, 00:12:38.408 "data_size": 65536 00:12:38.408 }, 00:12:38.408 { 00:12:38.408 "name": "BaseBdev3", 00:12:38.408 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:38.408 "is_configured": true, 00:12:38.408 "data_offset": 0, 00:12:38.408 "data_size": 65536 00:12:38.408 }, 00:12:38.408 { 00:12:38.408 "name": "BaseBdev4", 00:12:38.408 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:38.408 "is_configured": true, 00:12:38.408 "data_offset": 0, 00:12:38.408 "data_size": 65536 00:12:38.408 } 00:12:38.408 ] 00:12:38.408 }' 00:12:38.408 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:38.408 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:38.408 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:38.408 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:38.408 16:50:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:38.408 16:50:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.408 16:50:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.408 [2024-09-29 16:50:59.916302] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:38.408 [2024-09-29 16:50:59.920333] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:38.408 [2024-09-29 16:51:00.054630] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:38.408 [2024-09-29 16:51:00.061039] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:38.408 [2024-09-29 16:51:00.064236] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:38.408 [2024-09-29 16:51:00.064328] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:38.408 [2024-09-29 16:51:00.064355] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:38.408 [2024-09-29 16:51:00.070212] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:38.668 "name": "raid_bdev1", 00:12:38.668 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:38.668 "strip_size_kb": 0, 00:12:38.668 "state": "online", 00:12:38.668 "raid_level": "raid1", 00:12:38.668 "superblock": false, 00:12:38.668 "num_base_bdevs": 4, 00:12:38.668 "num_base_bdevs_discovered": 3, 00:12:38.668 "num_base_bdevs_operational": 3, 00:12:38.668 "base_bdevs_list": [ 00:12:38.668 { 00:12:38.668 "name": null, 00:12:38.668 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.668 "is_configured": false, 00:12:38.668 "data_offset": 0, 00:12:38.668 "data_size": 65536 00:12:38.668 }, 00:12:38.668 { 00:12:38.668 "name": "BaseBdev2", 00:12:38.668 "uuid": "7fedfea9-fe9c-5c07-adab-f829c7027c43", 00:12:38.668 "is_configured": true, 00:12:38.668 "data_offset": 0, 00:12:38.668 "data_size": 65536 00:12:38.668 }, 00:12:38.668 { 00:12:38.668 "name": "BaseBdev3", 00:12:38.668 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:38.668 "is_configured": true, 00:12:38.668 "data_offset": 0, 00:12:38.668 "data_size": 65536 00:12:38.668 }, 00:12:38.668 { 00:12:38.668 "name": "BaseBdev4", 00:12:38.668 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:38.668 "is_configured": true, 00:12:38.668 "data_offset": 0, 00:12:38.668 "data_size": 65536 00:12:38.668 } 00:12:38.668 ] 00:12:38.668 }' 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:38.668 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.928 160.50 IOPS, 481.50 MiB/s 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:38.928 "name": "raid_bdev1", 00:12:38.928 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:38.928 "strip_size_kb": 0, 00:12:38.928 "state": "online", 00:12:38.928 "raid_level": "raid1", 00:12:38.928 "superblock": false, 00:12:38.928 "num_base_bdevs": 4, 00:12:38.928 "num_base_bdevs_discovered": 3, 00:12:38.928 "num_base_bdevs_operational": 3, 00:12:38.928 "base_bdevs_list": [ 00:12:38.928 { 00:12:38.928 "name": null, 00:12:38.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.928 "is_configured": false, 00:12:38.928 "data_offset": 0, 00:12:38.928 "data_size": 65536 00:12:38.928 }, 00:12:38.928 { 00:12:38.928 "name": "BaseBdev2", 00:12:38.928 "uuid": "7fedfea9-fe9c-5c07-adab-f829c7027c43", 00:12:38.928 "is_configured": true, 00:12:38.928 "data_offset": 0, 00:12:38.928 "data_size": 65536 00:12:38.928 }, 00:12:38.928 { 00:12:38.928 "name": "BaseBdev3", 00:12:38.928 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:38.928 "is_configured": true, 00:12:38.928 "data_offset": 0, 00:12:38.928 "data_size": 65536 00:12:38.928 }, 00:12:38.928 { 00:12:38.928 "name": "BaseBdev4", 00:12:38.928 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:38.928 "is_configured": true, 00:12:38.928 "data_offset": 0, 00:12:38.928 "data_size": 65536 00:12:38.928 } 00:12:38.928 ] 00:12:38.928 }' 00:12:38.928 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:39.188 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:39.188 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:39.188 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:39.188 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:39.188 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.189 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:39.189 [2024-09-29 16:51:00.693454] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:39.189 16:51:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.189 16:51:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:39.189 [2024-09-29 16:51:00.747679] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:39.189 [2024-09-29 16:51:00.749754] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:39.448 [2024-09-29 16:51:00.876065] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:39.448 [2024-09-29 16:51:00.876616] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:39.448 [2024-09-29 16:51:00.999887] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:39.448 [2024-09-29 16:51:01.000165] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:39.708 [2024-09-29 16:51:01.357102] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:39.708 [2024-09-29 16:51:01.357699] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:39.967 151.00 IOPS, 453.00 MiB/s [2024-09-29 16:51:01.565309] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:39.967 [2024-09-29 16:51:01.565606] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:40.227 "name": "raid_bdev1", 00:12:40.227 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:40.227 "strip_size_kb": 0, 00:12:40.227 "state": "online", 00:12:40.227 "raid_level": "raid1", 00:12:40.227 "superblock": false, 00:12:40.227 "num_base_bdevs": 4, 00:12:40.227 "num_base_bdevs_discovered": 4, 00:12:40.227 "num_base_bdevs_operational": 4, 00:12:40.227 "process": { 00:12:40.227 "type": "rebuild", 00:12:40.227 "target": "spare", 00:12:40.227 "progress": { 00:12:40.227 "blocks": 10240, 00:12:40.227 "percent": 15 00:12:40.227 } 00:12:40.227 }, 00:12:40.227 "base_bdevs_list": [ 00:12:40.227 { 00:12:40.227 "name": "spare", 00:12:40.227 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:40.227 "is_configured": true, 00:12:40.227 "data_offset": 0, 00:12:40.227 "data_size": 65536 00:12:40.227 }, 00:12:40.227 { 00:12:40.227 "name": "BaseBdev2", 00:12:40.227 "uuid": "7fedfea9-fe9c-5c07-adab-f829c7027c43", 00:12:40.227 "is_configured": true, 00:12:40.227 "data_offset": 0, 00:12:40.227 "data_size": 65536 00:12:40.227 }, 00:12:40.227 { 00:12:40.227 "name": "BaseBdev3", 00:12:40.227 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:40.227 "is_configured": true, 00:12:40.227 "data_offset": 0, 00:12:40.227 "data_size": 65536 00:12:40.227 }, 00:12:40.227 { 00:12:40.227 "name": "BaseBdev4", 00:12:40.227 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:40.227 "is_configured": true, 00:12:40.227 "data_offset": 0, 00:12:40.227 "data_size": 65536 00:12:40.227 } 00:12:40.227 ] 00:12:40.227 }' 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.227 16:51:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.227 [2024-09-29 16:51:01.885711] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:40.227 [2024-09-29 16:51:01.893989] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:40.227 [2024-09-29 16:51:01.895231] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:40.488 [2024-09-29 16:51:01.996968] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:12:40.488 [2024-09-29 16:51:01.997037] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:12:40.488 16:51:01 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.488 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:40.488 16:51:01 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:40.488 "name": "raid_bdev1", 00:12:40.488 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:40.488 "strip_size_kb": 0, 00:12:40.488 "state": "online", 00:12:40.488 "raid_level": "raid1", 00:12:40.488 "superblock": false, 00:12:40.488 "num_base_bdevs": 4, 00:12:40.488 "num_base_bdevs_discovered": 3, 00:12:40.488 "num_base_bdevs_operational": 3, 00:12:40.488 "process": { 00:12:40.488 "type": "rebuild", 00:12:40.488 "target": "spare", 00:12:40.488 "progress": { 00:12:40.488 "blocks": 14336, 00:12:40.488 "percent": 21 00:12:40.488 } 00:12:40.488 }, 00:12:40.488 "base_bdevs_list": [ 00:12:40.488 { 00:12:40.488 "name": "spare", 00:12:40.488 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:40.488 "is_configured": true, 00:12:40.488 "data_offset": 0, 00:12:40.488 "data_size": 65536 00:12:40.488 }, 00:12:40.488 { 00:12:40.488 "name": null, 00:12:40.488 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.488 "is_configured": false, 00:12:40.488 "data_offset": 0, 00:12:40.488 "data_size": 65536 00:12:40.488 }, 00:12:40.488 { 00:12:40.488 "name": "BaseBdev3", 00:12:40.488 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:40.488 "is_configured": true, 00:12:40.488 "data_offset": 0, 00:12:40.488 "data_size": 65536 00:12:40.488 }, 00:12:40.488 { 00:12:40.488 "name": "BaseBdev4", 00:12:40.488 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:40.488 "is_configured": true, 00:12:40.488 "data_offset": 0, 00:12:40.488 "data_size": 65536 00:12:40.488 } 00:12:40.488 ] 00:12:40.488 }' 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:40.488 [2024-09-29 16:51:02.127301] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:40.488 [2024-09-29 16:51:02.127821] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=387 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.488 16:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:40.748 16:51:02 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.748 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:40.748 "name": "raid_bdev1", 00:12:40.748 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:40.748 "strip_size_kb": 0, 00:12:40.748 "state": "online", 00:12:40.748 "raid_level": "raid1", 00:12:40.748 "superblock": false, 00:12:40.748 "num_base_bdevs": 4, 00:12:40.748 "num_base_bdevs_discovered": 3, 00:12:40.748 "num_base_bdevs_operational": 3, 00:12:40.748 "process": { 00:12:40.748 "type": "rebuild", 00:12:40.748 "target": "spare", 00:12:40.748 "progress": { 00:12:40.748 "blocks": 16384, 00:12:40.748 "percent": 25 00:12:40.748 } 00:12:40.748 }, 00:12:40.748 "base_bdevs_list": [ 00:12:40.748 { 00:12:40.748 "name": "spare", 00:12:40.749 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:40.749 "is_configured": true, 00:12:40.749 "data_offset": 0, 00:12:40.749 "data_size": 65536 00:12:40.749 }, 00:12:40.749 { 00:12:40.749 "name": null, 00:12:40.749 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:40.749 "is_configured": false, 00:12:40.749 "data_offset": 0, 00:12:40.749 "data_size": 65536 00:12:40.749 }, 00:12:40.749 { 00:12:40.749 "name": "BaseBdev3", 00:12:40.749 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:40.749 "is_configured": true, 00:12:40.749 "data_offset": 0, 00:12:40.749 "data_size": 65536 00:12:40.749 }, 00:12:40.749 { 00:12:40.749 "name": "BaseBdev4", 00:12:40.749 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:40.749 "is_configured": true, 00:12:40.749 "data_offset": 0, 00:12:40.749 "data_size": 65536 00:12:40.749 } 00:12:40.749 ] 00:12:40.749 }' 00:12:40.749 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:40.749 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:40.749 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:40.749 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:40.749 16:51:02 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:41.007 125.50 IOPS, 376.50 MiB/s [2024-09-29 16:51:02.480900] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:41.267 [2024-09-29 16:51:02.712654] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:41.527 [2024-09-29 16:51:03.024238] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:41.787 "name": "raid_bdev1", 00:12:41.787 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:41.787 "strip_size_kb": 0, 00:12:41.787 "state": "online", 00:12:41.787 "raid_level": "raid1", 00:12:41.787 "superblock": false, 00:12:41.787 "num_base_bdevs": 4, 00:12:41.787 "num_base_bdevs_discovered": 3, 00:12:41.787 "num_base_bdevs_operational": 3, 00:12:41.787 "process": { 00:12:41.787 "type": "rebuild", 00:12:41.787 "target": "spare", 00:12:41.787 "progress": { 00:12:41.787 "blocks": 30720, 00:12:41.787 "percent": 46 00:12:41.787 } 00:12:41.787 }, 00:12:41.787 "base_bdevs_list": [ 00:12:41.787 { 00:12:41.787 "name": "spare", 00:12:41.787 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:41.787 "is_configured": true, 00:12:41.787 "data_offset": 0, 00:12:41.787 "data_size": 65536 00:12:41.787 }, 00:12:41.787 { 00:12:41.787 "name": null, 00:12:41.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:41.787 "is_configured": false, 00:12:41.787 "data_offset": 0, 00:12:41.787 "data_size": 65536 00:12:41.787 }, 00:12:41.787 { 00:12:41.787 "name": "BaseBdev3", 00:12:41.787 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:41.787 "is_configured": true, 00:12:41.787 "data_offset": 0, 00:12:41.787 "data_size": 65536 00:12:41.787 }, 00:12:41.787 { 00:12:41.787 "name": "BaseBdev4", 00:12:41.787 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:41.787 "is_configured": true, 00:12:41.787 "data_offset": 0, 00:12:41.787 "data_size": 65536 00:12:41.787 } 00:12:41.787 ] 00:12:41.787 }' 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:41.787 [2024-09-29 16:51:03.387984] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:41.787 115.60 IOPS, 346.80 MiB/s 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:41.787 16:51:03 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:42.356 [2024-09-29 16:51:03.729340] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:12:42.356 [2024-09-29 16:51:03.730129] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:12:43.029 105.00 IOPS, 315.00 MiB/s 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:43.029 "name": "raid_bdev1", 00:12:43.029 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:43.029 "strip_size_kb": 0, 00:12:43.029 "state": "online", 00:12:43.029 "raid_level": "raid1", 00:12:43.029 "superblock": false, 00:12:43.029 "num_base_bdevs": 4, 00:12:43.029 "num_base_bdevs_discovered": 3, 00:12:43.029 "num_base_bdevs_operational": 3, 00:12:43.029 "process": { 00:12:43.029 "type": "rebuild", 00:12:43.029 "target": "spare", 00:12:43.029 "progress": { 00:12:43.029 "blocks": 49152, 00:12:43.029 "percent": 75 00:12:43.029 } 00:12:43.029 }, 00:12:43.029 "base_bdevs_list": [ 00:12:43.029 { 00:12:43.029 "name": "spare", 00:12:43.029 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:43.029 "is_configured": true, 00:12:43.029 "data_offset": 0, 00:12:43.029 "data_size": 65536 00:12:43.029 }, 00:12:43.029 { 00:12:43.029 "name": null, 00:12:43.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:43.029 "is_configured": false, 00:12:43.029 "data_offset": 0, 00:12:43.029 "data_size": 65536 00:12:43.029 }, 00:12:43.029 { 00:12:43.029 "name": "BaseBdev3", 00:12:43.029 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:43.029 "is_configured": true, 00:12:43.029 "data_offset": 0, 00:12:43.029 "data_size": 65536 00:12:43.029 }, 00:12:43.029 { 00:12:43.029 "name": "BaseBdev4", 00:12:43.029 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:43.029 "is_configured": true, 00:12:43.029 "data_offset": 0, 00:12:43.029 "data_size": 65536 00:12:43.029 } 00:12:43.029 ] 00:12:43.029 }' 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:43.029 16:51:04 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:43.029 [2024-09-29 16:51:04.566201] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:12:43.305 [2024-09-29 16:51:04.888877] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:12:43.565 [2024-09-29 16:51:05.214602] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:43.826 [2024-09-29 16:51:05.319359] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:43.826 [2024-09-29 16:51:05.321542] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:44.087 95.00 IOPS, 285.00 MiB/s 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.087 "name": "raid_bdev1", 00:12:44.087 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:44.087 "strip_size_kb": 0, 00:12:44.087 "state": "online", 00:12:44.087 "raid_level": "raid1", 00:12:44.087 "superblock": false, 00:12:44.087 "num_base_bdevs": 4, 00:12:44.087 "num_base_bdevs_discovered": 3, 00:12:44.087 "num_base_bdevs_operational": 3, 00:12:44.087 "base_bdevs_list": [ 00:12:44.087 { 00:12:44.087 "name": "spare", 00:12:44.087 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:44.087 "is_configured": true, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 }, 00:12:44.087 { 00:12:44.087 "name": null, 00:12:44.087 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.087 "is_configured": false, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 }, 00:12:44.087 { 00:12:44.087 "name": "BaseBdev3", 00:12:44.087 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:44.087 "is_configured": true, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 }, 00:12:44.087 { 00:12:44.087 "name": "BaseBdev4", 00:12:44.087 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:44.087 "is_configured": true, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 } 00:12:44.087 ] 00:12:44.087 }' 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:44.087 "name": "raid_bdev1", 00:12:44.087 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:44.087 "strip_size_kb": 0, 00:12:44.087 "state": "online", 00:12:44.087 "raid_level": "raid1", 00:12:44.087 "superblock": false, 00:12:44.087 "num_base_bdevs": 4, 00:12:44.087 "num_base_bdevs_discovered": 3, 00:12:44.087 "num_base_bdevs_operational": 3, 00:12:44.087 "base_bdevs_list": [ 00:12:44.087 { 00:12:44.087 "name": "spare", 00:12:44.087 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:44.087 "is_configured": true, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 }, 00:12:44.087 { 00:12:44.087 "name": null, 00:12:44.087 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.087 "is_configured": false, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 }, 00:12:44.087 { 00:12:44.087 "name": "BaseBdev3", 00:12:44.087 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:44.087 "is_configured": true, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 }, 00:12:44.087 { 00:12:44.087 "name": "BaseBdev4", 00:12:44.087 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:44.087 "is_configured": true, 00:12:44.087 "data_offset": 0, 00:12:44.087 "data_size": 65536 00:12:44.087 } 00:12:44.087 ] 00:12:44.087 }' 00:12:44.087 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:44.348 "name": "raid_bdev1", 00:12:44.348 "uuid": "eadc630f-bded-4114-8d62-2a820adb1029", 00:12:44.348 "strip_size_kb": 0, 00:12:44.348 "state": "online", 00:12:44.348 "raid_level": "raid1", 00:12:44.348 "superblock": false, 00:12:44.348 "num_base_bdevs": 4, 00:12:44.348 "num_base_bdevs_discovered": 3, 00:12:44.348 "num_base_bdevs_operational": 3, 00:12:44.348 "base_bdevs_list": [ 00:12:44.348 { 00:12:44.348 "name": "spare", 00:12:44.348 "uuid": "852cc8bc-aa02-52bc-b350-3c4bc0d94075", 00:12:44.348 "is_configured": true, 00:12:44.348 "data_offset": 0, 00:12:44.348 "data_size": 65536 00:12:44.348 }, 00:12:44.348 { 00:12:44.348 "name": null, 00:12:44.348 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:44.348 "is_configured": false, 00:12:44.348 "data_offset": 0, 00:12:44.348 "data_size": 65536 00:12:44.348 }, 00:12:44.348 { 00:12:44.348 "name": "BaseBdev3", 00:12:44.348 "uuid": "7e813b3e-0108-5ab3-bb47-e5bc85bbf2db", 00:12:44.348 "is_configured": true, 00:12:44.348 "data_offset": 0, 00:12:44.348 "data_size": 65536 00:12:44.348 }, 00:12:44.348 { 00:12:44.348 "name": "BaseBdev4", 00:12:44.348 "uuid": "d0e728c6-97ec-55c5-812b-52cebe0717bb", 00:12:44.348 "is_configured": true, 00:12:44.348 "data_offset": 0, 00:12:44.348 "data_size": 65536 00:12:44.348 } 00:12:44.348 ] 00:12:44.348 }' 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:44.348 16:51:05 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.919 [2024-09-29 16:51:06.299013] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:44.919 [2024-09-29 16:51:06.299100] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:44.919 87.88 IOPS, 263.62 MiB/s 00:12:44.919 Latency(us) 00:12:44.919 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:44.919 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:44.919 raid_bdev1 : 8.03 87.71 263.14 0.00 0.00 15064.49 273.66 115847.04 00:12:44.919 =================================================================================================================== 00:12:44.919 Total : 87.71 263.14 0.00 0.00 15064.49 273.66 115847.04 00:12:44.919 [2024-09-29 16:51:06.397928] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:44.919 [2024-09-29 16:51:06.397997] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:44.919 [2024-09-29 16:51:06.398131] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:44.919 [2024-09-29 16:51:06.398177] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:44.919 { 00:12:44.919 "results": [ 00:12:44.919 { 00:12:44.919 "job": "raid_bdev1", 00:12:44.919 "core_mask": "0x1", 00:12:44.919 "workload": "randrw", 00:12:44.919 "percentage": 50, 00:12:44.919 "status": "finished", 00:12:44.919 "queue_depth": 2, 00:12:44.919 "io_size": 3145728, 00:12:44.919 "runtime": 8.026274, 00:12:44.919 "iops": 87.71193208704312, 00:12:44.919 "mibps": 263.13579626112937, 00:12:44.919 "io_failed": 0, 00:12:44.919 "io_timeout": 0, 00:12:44.919 "avg_latency_us": 15064.488447796746, 00:12:44.919 "min_latency_us": 273.6628820960699, 00:12:44.919 "max_latency_us": 115847.04279475982 00:12:44.919 } 00:12:44.919 ], 00:12:44.919 "core_count": 1 00:12:44.919 } 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:44.919 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:45.180 /dev/nbd0 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:45.180 1+0 records in 00:12:45.180 1+0 records out 00:12:45.180 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000592875 s, 6.9 MB/s 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:45.180 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:12:45.440 /dev/nbd1 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:45.440 1+0 records in 00:12:45.440 1+0 records out 00:12:45.440 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000372874 s, 11.0 MB/s 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:45.440 16:51:06 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:45.440 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:45.440 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:45.440 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:45.440 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:45.440 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:45.440 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:45.440 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:45.700 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:45.701 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:12:45.961 /dev/nbd1 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:45.961 1+0 records in 00:12:45.961 1+0 records out 00:12:45.961 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000402773 s, 10.2 MB/s 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:45.961 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:46.220 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 88991 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 88991 ']' 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 88991 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:46.481 16:51:07 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88991 00:12:46.481 killing process with pid 88991 00:12:46.481 Received shutdown signal, test time was about 9.638166 seconds 00:12:46.481 00:12:46.481 Latency(us) 00:12:46.481 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:46.481 =================================================================================================================== 00:12:46.481 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:46.481 16:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:46.481 16:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:46.481 16:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88991' 00:12:46.481 16:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 88991 00:12:46.481 [2024-09-29 16:51:08.003560] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:46.481 16:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 88991 00:12:46.481 [2024-09-29 16:51:08.048785] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:46.742 00:12:46.742 real 0m11.656s 00:12:46.742 user 0m15.079s 00:12:46.742 sys 0m1.841s 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:46.742 ************************************ 00:12:46.742 END TEST raid_rebuild_test_io 00:12:46.742 ************************************ 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:46.742 16:51:08 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:12:46.742 16:51:08 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:46.742 16:51:08 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:46.742 16:51:08 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:46.742 ************************************ 00:12:46.742 START TEST raid_rebuild_test_sb_io 00:12:46.742 ************************************ 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true true true 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89383 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89383 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 89383 ']' 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:46.742 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:46.742 16:51:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.003 [2024-09-29 16:51:08.475166] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:12:47.003 [2024-09-29 16:51:08.475388] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89383 ] 00:12:47.003 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:47.003 Zero copy mechanism will not be used. 00:12:47.003 [2024-09-29 16:51:08.622214] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:47.003 [2024-09-29 16:51:08.669262] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:47.263 [2024-09-29 16:51:08.712309] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:47.263 [2024-09-29 16:51:08.712343] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 BaseBdev1_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 [2024-09-29 16:51:09.318765] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:47.842 [2024-09-29 16:51:09.318874] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:47.842 [2024-09-29 16:51:09.318917] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:47.842 [2024-09-29 16:51:09.318949] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:47.842 [2024-09-29 16:51:09.321120] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:47.842 [2024-09-29 16:51:09.321186] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:47.842 BaseBdev1 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 BaseBdev2_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 [2024-09-29 16:51:09.363769] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:47.842 [2024-09-29 16:51:09.363891] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:47.842 [2024-09-29 16:51:09.363948] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:47.842 [2024-09-29 16:51:09.363974] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:47.842 [2024-09-29 16:51:09.368642] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:47.842 [2024-09-29 16:51:09.368707] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:47.842 BaseBdev2 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 BaseBdev3_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 [2024-09-29 16:51:09.394741] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:47.842 [2024-09-29 16:51:09.394838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:47.842 [2024-09-29 16:51:09.394887] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:47.842 [2024-09-29 16:51:09.394927] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:47.842 [2024-09-29 16:51:09.396997] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:47.842 [2024-09-29 16:51:09.397061] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:47.842 BaseBdev3 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 BaseBdev4_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 [2024-09-29 16:51:09.423326] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:47.842 [2024-09-29 16:51:09.423428] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:47.842 [2024-09-29 16:51:09.423467] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:47.842 [2024-09-29 16:51:09.423493] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:47.842 [2024-09-29 16:51:09.425500] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:47.842 [2024-09-29 16:51:09.425567] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:47.842 BaseBdev4 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 spare_malloc 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 spare_delay 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 [2024-09-29 16:51:09.463941] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:47.842 [2024-09-29 16:51:09.464042] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:47.842 [2024-09-29 16:51:09.464077] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:47.842 [2024-09-29 16:51:09.464103] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:47.842 [2024-09-29 16:51:09.466116] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:47.842 [2024-09-29 16:51:09.466181] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:47.842 spare 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.842 [2024-09-29 16:51:09.475991] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:47.842 [2024-09-29 16:51:09.477834] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:47.842 [2024-09-29 16:51:09.477932] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:47.842 [2024-09-29 16:51:09.478015] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:47.842 [2024-09-29 16:51:09.478228] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:47.842 [2024-09-29 16:51:09.478275] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:47.842 [2024-09-29 16:51:09.478540] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:47.842 [2024-09-29 16:51:09.478711] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:47.842 [2024-09-29 16:51:09.478776] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:47.842 [2024-09-29 16:51:09.478971] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:47.842 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:47.843 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.103 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:48.103 "name": "raid_bdev1", 00:12:48.103 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:48.103 "strip_size_kb": 0, 00:12:48.103 "state": "online", 00:12:48.103 "raid_level": "raid1", 00:12:48.103 "superblock": true, 00:12:48.103 "num_base_bdevs": 4, 00:12:48.103 "num_base_bdevs_discovered": 4, 00:12:48.103 "num_base_bdevs_operational": 4, 00:12:48.103 "base_bdevs_list": [ 00:12:48.103 { 00:12:48.103 "name": "BaseBdev1", 00:12:48.103 "uuid": "06e23c07-d713-5d53-8ff7-fbdcf074a0f8", 00:12:48.103 "is_configured": true, 00:12:48.103 "data_offset": 2048, 00:12:48.103 "data_size": 63488 00:12:48.103 }, 00:12:48.103 { 00:12:48.103 "name": "BaseBdev2", 00:12:48.103 "uuid": "0bfc10ee-f983-52f8-876b-60192334b450", 00:12:48.103 "is_configured": true, 00:12:48.103 "data_offset": 2048, 00:12:48.103 "data_size": 63488 00:12:48.103 }, 00:12:48.103 { 00:12:48.103 "name": "BaseBdev3", 00:12:48.103 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:48.103 "is_configured": true, 00:12:48.103 "data_offset": 2048, 00:12:48.103 "data_size": 63488 00:12:48.103 }, 00:12:48.103 { 00:12:48.103 "name": "BaseBdev4", 00:12:48.103 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:48.103 "is_configured": true, 00:12:48.103 "data_offset": 2048, 00:12:48.103 "data_size": 63488 00:12:48.103 } 00:12:48.103 ] 00:12:48.103 }' 00:12:48.103 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:48.103 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.364 [2024-09-29 16:51:09.943464] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:48.364 16:51:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.364 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:48.364 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:48.364 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:48.364 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.624 [2024-09-29 16:51:10.043063] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:48.624 "name": "raid_bdev1", 00:12:48.624 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:48.624 "strip_size_kb": 0, 00:12:48.624 "state": "online", 00:12:48.624 "raid_level": "raid1", 00:12:48.624 "superblock": true, 00:12:48.624 "num_base_bdevs": 4, 00:12:48.624 "num_base_bdevs_discovered": 3, 00:12:48.624 "num_base_bdevs_operational": 3, 00:12:48.624 "base_bdevs_list": [ 00:12:48.624 { 00:12:48.624 "name": null, 00:12:48.624 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.624 "is_configured": false, 00:12:48.624 "data_offset": 0, 00:12:48.624 "data_size": 63488 00:12:48.624 }, 00:12:48.624 { 00:12:48.624 "name": "BaseBdev2", 00:12:48.624 "uuid": "0bfc10ee-f983-52f8-876b-60192334b450", 00:12:48.624 "is_configured": true, 00:12:48.624 "data_offset": 2048, 00:12:48.624 "data_size": 63488 00:12:48.624 }, 00:12:48.624 { 00:12:48.624 "name": "BaseBdev3", 00:12:48.624 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:48.624 "is_configured": true, 00:12:48.624 "data_offset": 2048, 00:12:48.624 "data_size": 63488 00:12:48.624 }, 00:12:48.624 { 00:12:48.624 "name": "BaseBdev4", 00:12:48.624 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:48.624 "is_configured": true, 00:12:48.624 "data_offset": 2048, 00:12:48.624 "data_size": 63488 00:12:48.624 } 00:12:48.624 ] 00:12:48.624 }' 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:48.624 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.624 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:48.624 Zero copy mechanism will not be used. 00:12:48.624 Running I/O for 60 seconds... 00:12:48.624 [2024-09-29 16:51:10.132884] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:48.884 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:48.884 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.884 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:48.884 [2024-09-29 16:51:10.476923] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:48.884 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.884 16:51:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:48.884 [2024-09-29 16:51:10.523389] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:12:48.884 [2024-09-29 16:51:10.525332] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:49.144 [2024-09-29 16:51:10.640138] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:49.144 [2024-09-29 16:51:10.640520] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:49.404 [2024-09-29 16:51:10.868839] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:49.404 [2024-09-29 16:51:10.869139] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:49.663 [2024-09-29 16:51:11.117880] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:49.663 [2024-09-29 16:51:11.118239] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:49.922 172.00 IOPS, 516.00 MiB/s [2024-09-29 16:51:11.348850] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.922 "name": "raid_bdev1", 00:12:49.922 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:49.922 "strip_size_kb": 0, 00:12:49.922 "state": "online", 00:12:49.922 "raid_level": "raid1", 00:12:49.922 "superblock": true, 00:12:49.922 "num_base_bdevs": 4, 00:12:49.922 "num_base_bdevs_discovered": 4, 00:12:49.922 "num_base_bdevs_operational": 4, 00:12:49.922 "process": { 00:12:49.922 "type": "rebuild", 00:12:49.922 "target": "spare", 00:12:49.922 "progress": { 00:12:49.922 "blocks": 10240, 00:12:49.922 "percent": 16 00:12:49.922 } 00:12:49.922 }, 00:12:49.922 "base_bdevs_list": [ 00:12:49.922 { 00:12:49.922 "name": "spare", 00:12:49.922 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:49.922 "is_configured": true, 00:12:49.922 "data_offset": 2048, 00:12:49.922 "data_size": 63488 00:12:49.922 }, 00:12:49.922 { 00:12:49.922 "name": "BaseBdev2", 00:12:49.922 "uuid": "0bfc10ee-f983-52f8-876b-60192334b450", 00:12:49.922 "is_configured": true, 00:12:49.922 "data_offset": 2048, 00:12:49.922 "data_size": 63488 00:12:49.922 }, 00:12:49.922 { 00:12:49.922 "name": "BaseBdev3", 00:12:49.922 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:49.922 "is_configured": true, 00:12:49.922 "data_offset": 2048, 00:12:49.922 "data_size": 63488 00:12:49.922 }, 00:12:49.922 { 00:12:49.922 "name": "BaseBdev4", 00:12:49.922 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:49.922 "is_configured": true, 00:12:49.922 "data_offset": 2048, 00:12:49.922 "data_size": 63488 00:12:49.922 } 00:12:49.922 ] 00:12:49.922 }' 00:12:49.922 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:50.182 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:50.182 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:50.182 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:50.182 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:50.182 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.182 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.182 [2024-09-29 16:51:11.662298] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:50.182 [2024-09-29 16:51:11.697012] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:50.182 [2024-09-29 16:51:11.708777] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:50.182 [2024-09-29 16:51:11.717355] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:50.182 [2024-09-29 16:51:11.717397] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:50.183 [2024-09-29 16:51:11.717408] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:50.183 [2024-09-29 16:51:11.745237] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:50.183 "name": "raid_bdev1", 00:12:50.183 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:50.183 "strip_size_kb": 0, 00:12:50.183 "state": "online", 00:12:50.183 "raid_level": "raid1", 00:12:50.183 "superblock": true, 00:12:50.183 "num_base_bdevs": 4, 00:12:50.183 "num_base_bdevs_discovered": 3, 00:12:50.183 "num_base_bdevs_operational": 3, 00:12:50.183 "base_bdevs_list": [ 00:12:50.183 { 00:12:50.183 "name": null, 00:12:50.183 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.183 "is_configured": false, 00:12:50.183 "data_offset": 0, 00:12:50.183 "data_size": 63488 00:12:50.183 }, 00:12:50.183 { 00:12:50.183 "name": "BaseBdev2", 00:12:50.183 "uuid": "0bfc10ee-f983-52f8-876b-60192334b450", 00:12:50.183 "is_configured": true, 00:12:50.183 "data_offset": 2048, 00:12:50.183 "data_size": 63488 00:12:50.183 }, 00:12:50.183 { 00:12:50.183 "name": "BaseBdev3", 00:12:50.183 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:50.183 "is_configured": true, 00:12:50.183 "data_offset": 2048, 00:12:50.183 "data_size": 63488 00:12:50.183 }, 00:12:50.183 { 00:12:50.183 "name": "BaseBdev4", 00:12:50.183 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:50.183 "is_configured": true, 00:12:50.183 "data_offset": 2048, 00:12:50.183 "data_size": 63488 00:12:50.183 } 00:12:50.183 ] 00:12:50.183 }' 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:50.183 16:51:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.753 164.50 IOPS, 493.50 MiB/s 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:50.753 "name": "raid_bdev1", 00:12:50.753 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:50.753 "strip_size_kb": 0, 00:12:50.753 "state": "online", 00:12:50.753 "raid_level": "raid1", 00:12:50.753 "superblock": true, 00:12:50.753 "num_base_bdevs": 4, 00:12:50.753 "num_base_bdevs_discovered": 3, 00:12:50.753 "num_base_bdevs_operational": 3, 00:12:50.753 "base_bdevs_list": [ 00:12:50.753 { 00:12:50.753 "name": null, 00:12:50.753 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.753 "is_configured": false, 00:12:50.753 "data_offset": 0, 00:12:50.753 "data_size": 63488 00:12:50.753 }, 00:12:50.753 { 00:12:50.753 "name": "BaseBdev2", 00:12:50.753 "uuid": "0bfc10ee-f983-52f8-876b-60192334b450", 00:12:50.753 "is_configured": true, 00:12:50.753 "data_offset": 2048, 00:12:50.753 "data_size": 63488 00:12:50.753 }, 00:12:50.753 { 00:12:50.753 "name": "BaseBdev3", 00:12:50.753 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:50.753 "is_configured": true, 00:12:50.753 "data_offset": 2048, 00:12:50.753 "data_size": 63488 00:12:50.753 }, 00:12:50.753 { 00:12:50.753 "name": "BaseBdev4", 00:12:50.753 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:50.753 "is_configured": true, 00:12:50.753 "data_offset": 2048, 00:12:50.753 "data_size": 63488 00:12:50.753 } 00:12:50.753 ] 00:12:50.753 }' 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:50.753 [2024-09-29 16:51:12.374524] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.753 16:51:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:50.753 [2024-09-29 16:51:12.403230] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:12:50.753 [2024-09-29 16:51:12.405333] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:51.012 [2024-09-29 16:51:12.526982] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:51.012 [2024-09-29 16:51:12.528310] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:51.271 [2024-09-29 16:51:12.738575] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:51.271 [2024-09-29 16:51:12.739000] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:51.530 [2024-09-29 16:51:13.067051] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:51.530 [2024-09-29 16:51:13.067576] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:51.530 171.00 IOPS, 513.00 MiB/s [2024-09-29 16:51:13.187355] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:51.789 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:51.789 "name": "raid_bdev1", 00:12:51.789 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:51.789 "strip_size_kb": 0, 00:12:51.789 "state": "online", 00:12:51.789 "raid_level": "raid1", 00:12:51.789 "superblock": true, 00:12:51.789 "num_base_bdevs": 4, 00:12:51.789 "num_base_bdevs_discovered": 4, 00:12:51.789 "num_base_bdevs_operational": 4, 00:12:51.789 "process": { 00:12:51.789 "type": "rebuild", 00:12:51.789 "target": "spare", 00:12:51.789 "progress": { 00:12:51.789 "blocks": 10240, 00:12:51.789 "percent": 16 00:12:51.789 } 00:12:51.789 }, 00:12:51.789 "base_bdevs_list": [ 00:12:51.789 { 00:12:51.789 "name": "spare", 00:12:51.789 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:51.789 "is_configured": true, 00:12:51.789 "data_offset": 2048, 00:12:51.789 "data_size": 63488 00:12:51.789 }, 00:12:51.789 { 00:12:51.789 "name": "BaseBdev2", 00:12:51.789 "uuid": "0bfc10ee-f983-52f8-876b-60192334b450", 00:12:51.789 "is_configured": true, 00:12:51.789 "data_offset": 2048, 00:12:51.790 "data_size": 63488 00:12:51.790 }, 00:12:51.790 { 00:12:51.790 "name": "BaseBdev3", 00:12:51.790 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:51.790 "is_configured": true, 00:12:51.790 "data_offset": 2048, 00:12:51.790 "data_size": 63488 00:12:51.790 }, 00:12:51.790 { 00:12:51.790 "name": "BaseBdev4", 00:12:51.790 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:51.790 "is_configured": true, 00:12:51.790 "data_offset": 2048, 00:12:51.790 "data_size": 63488 00:12:51.790 } 00:12:51.790 ] 00:12:51.790 }' 00:12:51.790 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:52.049 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:52.049 [2024-09-29 16:51:13.547545] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.049 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.049 [2024-09-29 16:51:13.557603] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:52.307 [2024-09-29 16:51:13.750664] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:52.307 [2024-09-29 16:51:13.751395] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:52.307 [2024-09-29 16:51:13.958046] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:12:52.307 [2024-09-29 16:51:13.958120] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.307 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.567 16:51:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:52.567 "name": "raid_bdev1", 00:12:52.567 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:52.567 "strip_size_kb": 0, 00:12:52.567 "state": "online", 00:12:52.567 "raid_level": "raid1", 00:12:52.567 "superblock": true, 00:12:52.567 "num_base_bdevs": 4, 00:12:52.567 "num_base_bdevs_discovered": 3, 00:12:52.567 "num_base_bdevs_operational": 3, 00:12:52.567 "process": { 00:12:52.567 "type": "rebuild", 00:12:52.567 "target": "spare", 00:12:52.567 "progress": { 00:12:52.567 "blocks": 16384, 00:12:52.567 "percent": 25 00:12:52.567 } 00:12:52.567 }, 00:12:52.567 "base_bdevs_list": [ 00:12:52.567 { 00:12:52.567 "name": "spare", 00:12:52.567 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:52.567 "is_configured": true, 00:12:52.567 "data_offset": 2048, 00:12:52.567 "data_size": 63488 00:12:52.567 }, 00:12:52.567 { 00:12:52.567 "name": null, 00:12:52.567 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.567 "is_configured": false, 00:12:52.567 "data_offset": 0, 00:12:52.567 "data_size": 63488 00:12:52.567 }, 00:12:52.567 { 00:12:52.567 "name": "BaseBdev3", 00:12:52.567 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:52.567 "is_configured": true, 00:12:52.567 "data_offset": 2048, 00:12:52.567 "data_size": 63488 00:12:52.567 }, 00:12:52.567 { 00:12:52.567 "name": "BaseBdev4", 00:12:52.567 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:52.567 "is_configured": true, 00:12:52.567 "data_offset": 2048, 00:12:52.567 "data_size": 63488 00:12:52.567 } 00:12:52.567 ] 00:12:52.567 }' 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=399 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:52.567 146.75 IOPS, 440.25 MiB/s 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:52.567 "name": "raid_bdev1", 00:12:52.567 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:52.567 "strip_size_kb": 0, 00:12:52.567 "state": "online", 00:12:52.567 "raid_level": "raid1", 00:12:52.567 "superblock": true, 00:12:52.567 "num_base_bdevs": 4, 00:12:52.567 "num_base_bdevs_discovered": 3, 00:12:52.567 "num_base_bdevs_operational": 3, 00:12:52.567 "process": { 00:12:52.567 "type": "rebuild", 00:12:52.567 "target": "spare", 00:12:52.567 "progress": { 00:12:52.567 "blocks": 18432, 00:12:52.567 "percent": 29 00:12:52.567 } 00:12:52.567 }, 00:12:52.567 "base_bdevs_list": [ 00:12:52.567 { 00:12:52.567 "name": "spare", 00:12:52.567 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:52.567 "is_configured": true, 00:12:52.567 "data_offset": 2048, 00:12:52.567 "data_size": 63488 00:12:52.567 }, 00:12:52.567 { 00:12:52.567 "name": null, 00:12:52.567 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.567 "is_configured": false, 00:12:52.567 "data_offset": 0, 00:12:52.567 "data_size": 63488 00:12:52.567 }, 00:12:52.567 { 00:12:52.567 "name": "BaseBdev3", 00:12:52.567 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:52.567 "is_configured": true, 00:12:52.567 "data_offset": 2048, 00:12:52.567 "data_size": 63488 00:12:52.567 }, 00:12:52.567 { 00:12:52.567 "name": "BaseBdev4", 00:12:52.567 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:52.567 "is_configured": true, 00:12:52.567 "data_offset": 2048, 00:12:52.567 "data_size": 63488 00:12:52.567 } 00:12:52.567 ] 00:12:52.567 }' 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:52.567 [2024-09-29 16:51:14.184857] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:52.567 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:52.826 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:52.826 16:51:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:52.826 [2024-09-29 16:51:14.407294] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:53.086 [2024-09-29 16:51:14.754310] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:53.345 [2024-09-29 16:51:14.871750] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:53.604 128.20 IOPS, 384.60 MiB/s [2024-09-29 16:51:15.192623] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:53.604 [2024-09-29 16:51:15.192954] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:53.604 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:53.863 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:53.863 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:53.863 "name": "raid_bdev1", 00:12:53.863 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:53.863 "strip_size_kb": 0, 00:12:53.863 "state": "online", 00:12:53.863 "raid_level": "raid1", 00:12:53.863 "superblock": true, 00:12:53.863 "num_base_bdevs": 4, 00:12:53.863 "num_base_bdevs_discovered": 3, 00:12:53.863 "num_base_bdevs_operational": 3, 00:12:53.863 "process": { 00:12:53.863 "type": "rebuild", 00:12:53.863 "target": "spare", 00:12:53.863 "progress": { 00:12:53.863 "blocks": 32768, 00:12:53.863 "percent": 51 00:12:53.863 } 00:12:53.863 }, 00:12:53.863 "base_bdevs_list": [ 00:12:53.863 { 00:12:53.863 "name": "spare", 00:12:53.863 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:53.863 "is_configured": true, 00:12:53.863 "data_offset": 2048, 00:12:53.863 "data_size": 63488 00:12:53.863 }, 00:12:53.863 { 00:12:53.863 "name": null, 00:12:53.863 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:53.863 "is_configured": false, 00:12:53.863 "data_offset": 0, 00:12:53.863 "data_size": 63488 00:12:53.863 }, 00:12:53.863 { 00:12:53.863 "name": "BaseBdev3", 00:12:53.864 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:53.864 "is_configured": true, 00:12:53.864 "data_offset": 2048, 00:12:53.864 "data_size": 63488 00:12:53.864 }, 00:12:53.864 { 00:12:53.864 "name": "BaseBdev4", 00:12:53.864 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:53.864 "is_configured": true, 00:12:53.864 "data_offset": 2048, 00:12:53.864 "data_size": 63488 00:12:53.864 } 00:12:53.864 ] 00:12:53.864 }' 00:12:53.864 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:53.864 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:53.864 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:53.864 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:53.864 16:51:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:53.864 [2024-09-29 16:51:15.409627] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:54.802 114.83 IOPS, 344.50 MiB/s 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:54.802 "name": "raid_bdev1", 00:12:54.802 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:54.802 "strip_size_kb": 0, 00:12:54.802 "state": "online", 00:12:54.802 "raid_level": "raid1", 00:12:54.802 "superblock": true, 00:12:54.802 "num_base_bdevs": 4, 00:12:54.802 "num_base_bdevs_discovered": 3, 00:12:54.802 "num_base_bdevs_operational": 3, 00:12:54.802 "process": { 00:12:54.802 "type": "rebuild", 00:12:54.802 "target": "spare", 00:12:54.802 "progress": { 00:12:54.802 "blocks": 51200, 00:12:54.802 "percent": 80 00:12:54.802 } 00:12:54.802 }, 00:12:54.802 "base_bdevs_list": [ 00:12:54.802 { 00:12:54.802 "name": "spare", 00:12:54.802 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:54.802 "is_configured": true, 00:12:54.802 "data_offset": 2048, 00:12:54.802 "data_size": 63488 00:12:54.802 }, 00:12:54.802 { 00:12:54.802 "name": null, 00:12:54.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:54.802 "is_configured": false, 00:12:54.802 "data_offset": 0, 00:12:54.802 "data_size": 63488 00:12:54.802 }, 00:12:54.802 { 00:12:54.802 "name": "BaseBdev3", 00:12:54.802 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:54.802 "is_configured": true, 00:12:54.802 "data_offset": 2048, 00:12:54.802 "data_size": 63488 00:12:54.802 }, 00:12:54.802 { 00:12:54.802 "name": "BaseBdev4", 00:12:54.802 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:54.802 "is_configured": true, 00:12:54.802 "data_offset": 2048, 00:12:54.802 "data_size": 63488 00:12:54.802 } 00:12:54.802 ] 00:12:54.802 }' 00:12:54.802 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:55.062 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:55.062 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:55.062 [2024-09-29 16:51:16.526159] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:12:55.062 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:55.062 16:51:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:55.633 103.71 IOPS, 311.14 MiB/s [2024-09-29 16:51:17.182393] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:55.633 [2024-09-29 16:51:17.282212] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:55.633 [2024-09-29 16:51:17.284493] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:55.892 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:55.892 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:55.892 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:55.892 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:55.892 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:55.892 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:56.152 "name": "raid_bdev1", 00:12:56.152 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:56.152 "strip_size_kb": 0, 00:12:56.152 "state": "online", 00:12:56.152 "raid_level": "raid1", 00:12:56.152 "superblock": true, 00:12:56.152 "num_base_bdevs": 4, 00:12:56.152 "num_base_bdevs_discovered": 3, 00:12:56.152 "num_base_bdevs_operational": 3, 00:12:56.152 "base_bdevs_list": [ 00:12:56.152 { 00:12:56.152 "name": "spare", 00:12:56.152 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:56.152 "is_configured": true, 00:12:56.152 "data_offset": 2048, 00:12:56.152 "data_size": 63488 00:12:56.152 }, 00:12:56.152 { 00:12:56.152 "name": null, 00:12:56.152 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:56.152 "is_configured": false, 00:12:56.152 "data_offset": 0, 00:12:56.152 "data_size": 63488 00:12:56.152 }, 00:12:56.152 { 00:12:56.152 "name": "BaseBdev3", 00:12:56.152 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:56.152 "is_configured": true, 00:12:56.152 "data_offset": 2048, 00:12:56.152 "data_size": 63488 00:12:56.152 }, 00:12:56.152 { 00:12:56.152 "name": "BaseBdev4", 00:12:56.152 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:56.152 "is_configured": true, 00:12:56.152 "data_offset": 2048, 00:12:56.152 "data_size": 63488 00:12:56.152 } 00:12:56.152 ] 00:12:56.152 }' 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:56.152 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:56.152 "name": "raid_bdev1", 00:12:56.152 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:56.152 "strip_size_kb": 0, 00:12:56.152 "state": "online", 00:12:56.152 "raid_level": "raid1", 00:12:56.152 "superblock": true, 00:12:56.152 "num_base_bdevs": 4, 00:12:56.152 "num_base_bdevs_discovered": 3, 00:12:56.153 "num_base_bdevs_operational": 3, 00:12:56.153 "base_bdevs_list": [ 00:12:56.153 { 00:12:56.153 "name": "spare", 00:12:56.153 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:56.153 "is_configured": true, 00:12:56.153 "data_offset": 2048, 00:12:56.153 "data_size": 63488 00:12:56.153 }, 00:12:56.153 { 00:12:56.153 "name": null, 00:12:56.153 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:56.153 "is_configured": false, 00:12:56.153 "data_offset": 0, 00:12:56.153 "data_size": 63488 00:12:56.153 }, 00:12:56.153 { 00:12:56.153 "name": "BaseBdev3", 00:12:56.153 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:56.153 "is_configured": true, 00:12:56.153 "data_offset": 2048, 00:12:56.153 "data_size": 63488 00:12:56.153 }, 00:12:56.153 { 00:12:56.153 "name": "BaseBdev4", 00:12:56.153 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:56.153 "is_configured": true, 00:12:56.153 "data_offset": 2048, 00:12:56.153 "data_size": 63488 00:12:56.153 } 00:12:56.153 ] 00:12:56.153 }' 00:12:56.153 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:56.153 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:56.153 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:56.413 "name": "raid_bdev1", 00:12:56.413 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:56.413 "strip_size_kb": 0, 00:12:56.413 "state": "online", 00:12:56.413 "raid_level": "raid1", 00:12:56.413 "superblock": true, 00:12:56.413 "num_base_bdevs": 4, 00:12:56.413 "num_base_bdevs_discovered": 3, 00:12:56.413 "num_base_bdevs_operational": 3, 00:12:56.413 "base_bdevs_list": [ 00:12:56.413 { 00:12:56.413 "name": "spare", 00:12:56.413 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:56.413 "is_configured": true, 00:12:56.413 "data_offset": 2048, 00:12:56.413 "data_size": 63488 00:12:56.413 }, 00:12:56.413 { 00:12:56.413 "name": null, 00:12:56.413 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:56.413 "is_configured": false, 00:12:56.413 "data_offset": 0, 00:12:56.413 "data_size": 63488 00:12:56.413 }, 00:12:56.413 { 00:12:56.413 "name": "BaseBdev3", 00:12:56.413 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:56.413 "is_configured": true, 00:12:56.413 "data_offset": 2048, 00:12:56.413 "data_size": 63488 00:12:56.413 }, 00:12:56.413 { 00:12:56.413 "name": "BaseBdev4", 00:12:56.413 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:56.413 "is_configured": true, 00:12:56.413 "data_offset": 2048, 00:12:56.413 "data_size": 63488 00:12:56.413 } 00:12:56.413 ] 00:12:56.413 }' 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:56.413 16:51:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.673 94.12 IOPS, 282.38 MiB/s 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:56.673 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:56.673 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.673 [2024-09-29 16:51:18.286097] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:56.673 [2024-09-29 16:51:18.286167] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:56.934 00:12:56.934 Latency(us) 00:12:56.934 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:56.934 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:56.934 raid_bdev1 : 8.23 92.66 277.97 0.00 0.00 14357.21 273.66 115847.04 00:12:56.934 =================================================================================================================== 00:12:56.934 Total : 92.66 277.97 0.00 0.00 14357.21 273.66 115847.04 00:12:56.934 [2024-09-29 16:51:18.357096] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:56.934 [2024-09-29 16:51:18.357168] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:56.934 [2024-09-29 16:51:18.357310] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:56.934 [2024-09-29 16:51:18.357356] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:56.934 { 00:12:56.934 "results": [ 00:12:56.934 { 00:12:56.934 "job": "raid_bdev1", 00:12:56.934 "core_mask": "0x1", 00:12:56.934 "workload": "randrw", 00:12:56.934 "percentage": 50, 00:12:56.934 "status": "finished", 00:12:56.934 "queue_depth": 2, 00:12:56.934 "io_size": 3145728, 00:12:56.934 "runtime": 8.234811, 00:12:56.934 "iops": 92.65543556494497, 00:12:56.934 "mibps": 277.9663066948349, 00:12:56.934 "io_failed": 0, 00:12:56.934 "io_timeout": 0, 00:12:56.934 "avg_latency_us": 14357.207062445988, 00:12:56.934 "min_latency_us": 273.6628820960699, 00:12:56.934 "max_latency_us": 115847.04279475982 00:12:56.934 } 00:12:56.934 ], 00:12:56.934 "core_count": 1 00:12:56.934 } 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:56.934 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:57.194 /dev/nbd0 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:57.194 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:57.195 1+0 records in 00:12:57.195 1+0 records out 00:12:57.195 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000448438 s, 9.1 MB/s 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:57.195 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:12:57.455 /dev/nbd1 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:57.455 1+0 records in 00:12:57.455 1+0 records out 00:12:57.455 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000407004 s, 10.1 MB/s 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:57.455 16:51:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:57.455 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:57.455 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:57.455 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:57.455 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:57.455 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:57.455 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:57.455 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:57.715 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:12:57.975 /dev/nbd1 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:57.975 1+0 records in 00:12:57.975 1+0 records out 00:12:57.975 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000368185 s, 11.1 MB/s 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:57.975 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:58.235 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:58.495 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.495 [2024-09-29 16:51:19.995599] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:58.495 [2024-09-29 16:51:19.995700] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:58.495 [2024-09-29 16:51:19.995778] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:12:58.496 [2024-09-29 16:51:19.995811] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:58.496 [2024-09-29 16:51:19.998030] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:58.496 [2024-09-29 16:51:19.998117] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:58.496 [2024-09-29 16:51:19.998226] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:58.496 [2024-09-29 16:51:19.998292] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:58.496 [2024-09-29 16:51:19.998446] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:58.496 [2024-09-29 16:51:19.998578] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:58.496 spare 00:12:58.496 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:58.496 16:51:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.496 [2024-09-29 16:51:20.098532] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:58.496 [2024-09-29 16:51:20.098592] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:58.496 [2024-09-29 16:51:20.098892] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000337b0 00:12:58.496 [2024-09-29 16:51:20.099110] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:58.496 [2024-09-29 16:51:20.099159] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:58.496 [2024-09-29 16:51:20.099349] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:58.496 "name": "raid_bdev1", 00:12:58.496 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:58.496 "strip_size_kb": 0, 00:12:58.496 "state": "online", 00:12:58.496 "raid_level": "raid1", 00:12:58.496 "superblock": true, 00:12:58.496 "num_base_bdevs": 4, 00:12:58.496 "num_base_bdevs_discovered": 3, 00:12:58.496 "num_base_bdevs_operational": 3, 00:12:58.496 "base_bdevs_list": [ 00:12:58.496 { 00:12:58.496 "name": "spare", 00:12:58.496 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:58.496 "is_configured": true, 00:12:58.496 "data_offset": 2048, 00:12:58.496 "data_size": 63488 00:12:58.496 }, 00:12:58.496 { 00:12:58.496 "name": null, 00:12:58.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:58.496 "is_configured": false, 00:12:58.496 "data_offset": 2048, 00:12:58.496 "data_size": 63488 00:12:58.496 }, 00:12:58.496 { 00:12:58.496 "name": "BaseBdev3", 00:12:58.496 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:58.496 "is_configured": true, 00:12:58.496 "data_offset": 2048, 00:12:58.496 "data_size": 63488 00:12:58.496 }, 00:12:58.496 { 00:12:58.496 "name": "BaseBdev4", 00:12:58.496 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:58.496 "is_configured": true, 00:12:58.496 "data_offset": 2048, 00:12:58.496 "data_size": 63488 00:12:58.496 } 00:12:58.496 ] 00:12:58.496 }' 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:58.496 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:59.067 "name": "raid_bdev1", 00:12:59.067 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:59.067 "strip_size_kb": 0, 00:12:59.067 "state": "online", 00:12:59.067 "raid_level": "raid1", 00:12:59.067 "superblock": true, 00:12:59.067 "num_base_bdevs": 4, 00:12:59.067 "num_base_bdevs_discovered": 3, 00:12:59.067 "num_base_bdevs_operational": 3, 00:12:59.067 "base_bdevs_list": [ 00:12:59.067 { 00:12:59.067 "name": "spare", 00:12:59.067 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:12:59.067 "is_configured": true, 00:12:59.067 "data_offset": 2048, 00:12:59.067 "data_size": 63488 00:12:59.067 }, 00:12:59.067 { 00:12:59.067 "name": null, 00:12:59.067 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.067 "is_configured": false, 00:12:59.067 "data_offset": 2048, 00:12:59.067 "data_size": 63488 00:12:59.067 }, 00:12:59.067 { 00:12:59.067 "name": "BaseBdev3", 00:12:59.067 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:59.067 "is_configured": true, 00:12:59.067 "data_offset": 2048, 00:12:59.067 "data_size": 63488 00:12:59.067 }, 00:12:59.067 { 00:12:59.067 "name": "BaseBdev4", 00:12:59.067 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:59.067 "is_configured": true, 00:12:59.067 "data_offset": 2048, 00:12:59.067 "data_size": 63488 00:12:59.067 } 00:12:59.067 ] 00:12:59.067 }' 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.067 [2024-09-29 16:51:20.722501] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.067 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.327 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.327 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:59.327 "name": "raid_bdev1", 00:12:59.327 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:12:59.327 "strip_size_kb": 0, 00:12:59.327 "state": "online", 00:12:59.327 "raid_level": "raid1", 00:12:59.327 "superblock": true, 00:12:59.327 "num_base_bdevs": 4, 00:12:59.327 "num_base_bdevs_discovered": 2, 00:12:59.327 "num_base_bdevs_operational": 2, 00:12:59.327 "base_bdevs_list": [ 00:12:59.327 { 00:12:59.327 "name": null, 00:12:59.327 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.327 "is_configured": false, 00:12:59.327 "data_offset": 0, 00:12:59.327 "data_size": 63488 00:12:59.327 }, 00:12:59.327 { 00:12:59.327 "name": null, 00:12:59.327 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:59.327 "is_configured": false, 00:12:59.327 "data_offset": 2048, 00:12:59.327 "data_size": 63488 00:12:59.327 }, 00:12:59.327 { 00:12:59.327 "name": "BaseBdev3", 00:12:59.327 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:12:59.327 "is_configured": true, 00:12:59.327 "data_offset": 2048, 00:12:59.327 "data_size": 63488 00:12:59.327 }, 00:12:59.327 { 00:12:59.327 "name": "BaseBdev4", 00:12:59.327 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:12:59.327 "is_configured": true, 00:12:59.327 "data_offset": 2048, 00:12:59.327 "data_size": 63488 00:12:59.327 } 00:12:59.327 ] 00:12:59.327 }' 00:12:59.327 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:59.327 16:51:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.587 16:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:59.587 16:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:59.587 16:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:59.587 [2024-09-29 16:51:21.153842] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:59.587 [2024-09-29 16:51:21.154046] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:12:59.587 [2024-09-29 16:51:21.154106] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:59.587 [2024-09-29 16:51:21.154170] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:59.587 [2024-09-29 16:51:21.157856] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033880 00:12:59.587 16:51:21 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:59.587 16:51:21 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:59.587 [2024-09-29 16:51:21.159766] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:00.526 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:00.526 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:00.526 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:00.526 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:00.526 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:00.527 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.527 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.527 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.527 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.527 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:00.787 "name": "raid_bdev1", 00:13:00.787 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:00.787 "strip_size_kb": 0, 00:13:00.787 "state": "online", 00:13:00.787 "raid_level": "raid1", 00:13:00.787 "superblock": true, 00:13:00.787 "num_base_bdevs": 4, 00:13:00.787 "num_base_bdevs_discovered": 3, 00:13:00.787 "num_base_bdevs_operational": 3, 00:13:00.787 "process": { 00:13:00.787 "type": "rebuild", 00:13:00.787 "target": "spare", 00:13:00.787 "progress": { 00:13:00.787 "blocks": 20480, 00:13:00.787 "percent": 32 00:13:00.787 } 00:13:00.787 }, 00:13:00.787 "base_bdevs_list": [ 00:13:00.787 { 00:13:00.787 "name": "spare", 00:13:00.787 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:13:00.787 "is_configured": true, 00:13:00.787 "data_offset": 2048, 00:13:00.787 "data_size": 63488 00:13:00.787 }, 00:13:00.787 { 00:13:00.787 "name": null, 00:13:00.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.787 "is_configured": false, 00:13:00.787 "data_offset": 2048, 00:13:00.787 "data_size": 63488 00:13:00.787 }, 00:13:00.787 { 00:13:00.787 "name": "BaseBdev3", 00:13:00.787 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:00.787 "is_configured": true, 00:13:00.787 "data_offset": 2048, 00:13:00.787 "data_size": 63488 00:13:00.787 }, 00:13:00.787 { 00:13:00.787 "name": "BaseBdev4", 00:13:00.787 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:00.787 "is_configured": true, 00:13:00.787 "data_offset": 2048, 00:13:00.787 "data_size": 63488 00:13:00.787 } 00:13:00.787 ] 00:13:00.787 }' 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.787 [2024-09-29 16:51:22.324715] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:00.787 [2024-09-29 16:51:22.363761] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:00.787 [2024-09-29 16:51:22.363815] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:00.787 [2024-09-29 16:51:22.363834] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:00.787 [2024-09-29 16:51:22.363841] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:00.787 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:00.787 "name": "raid_bdev1", 00:13:00.787 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:00.787 "strip_size_kb": 0, 00:13:00.787 "state": "online", 00:13:00.787 "raid_level": "raid1", 00:13:00.787 "superblock": true, 00:13:00.787 "num_base_bdevs": 4, 00:13:00.787 "num_base_bdevs_discovered": 2, 00:13:00.787 "num_base_bdevs_operational": 2, 00:13:00.787 "base_bdevs_list": [ 00:13:00.787 { 00:13:00.787 "name": null, 00:13:00.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.787 "is_configured": false, 00:13:00.787 "data_offset": 0, 00:13:00.787 "data_size": 63488 00:13:00.787 }, 00:13:00.787 { 00:13:00.787 "name": null, 00:13:00.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:00.787 "is_configured": false, 00:13:00.788 "data_offset": 2048, 00:13:00.788 "data_size": 63488 00:13:00.788 }, 00:13:00.788 { 00:13:00.788 "name": "BaseBdev3", 00:13:00.788 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:00.788 "is_configured": true, 00:13:00.788 "data_offset": 2048, 00:13:00.788 "data_size": 63488 00:13:00.788 }, 00:13:00.788 { 00:13:00.788 "name": "BaseBdev4", 00:13:00.788 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:00.788 "is_configured": true, 00:13:00.788 "data_offset": 2048, 00:13:00.788 "data_size": 63488 00:13:00.788 } 00:13:00.788 ] 00:13:00.788 }' 00:13:00.788 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:00.788 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.357 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:01.358 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.358 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:01.358 [2024-09-29 16:51:22.842926] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:01.358 [2024-09-29 16:51:22.843042] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:01.358 [2024-09-29 16:51:22.843095] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:13:01.358 [2024-09-29 16:51:22.843125] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:01.358 [2024-09-29 16:51:22.843588] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:01.358 [2024-09-29 16:51:22.843644] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:01.358 [2024-09-29 16:51:22.843776] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:01.358 [2024-09-29 16:51:22.843817] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:01.358 [2024-09-29 16:51:22.843877] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:01.358 [2024-09-29 16:51:22.843924] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:01.358 [2024-09-29 16:51:22.847210] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033950 00:13:01.358 spare 00:13:01.358 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.358 16:51:22 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:01.358 [2024-09-29 16:51:22.849112] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:02.297 "name": "raid_bdev1", 00:13:02.297 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:02.297 "strip_size_kb": 0, 00:13:02.297 "state": "online", 00:13:02.297 "raid_level": "raid1", 00:13:02.297 "superblock": true, 00:13:02.297 "num_base_bdevs": 4, 00:13:02.297 "num_base_bdevs_discovered": 3, 00:13:02.297 "num_base_bdevs_operational": 3, 00:13:02.297 "process": { 00:13:02.297 "type": "rebuild", 00:13:02.297 "target": "spare", 00:13:02.297 "progress": { 00:13:02.297 "blocks": 20480, 00:13:02.297 "percent": 32 00:13:02.297 } 00:13:02.297 }, 00:13:02.297 "base_bdevs_list": [ 00:13:02.297 { 00:13:02.297 "name": "spare", 00:13:02.297 "uuid": "7b367fd7-dcc9-5b52-bf5b-3e8291a190b2", 00:13:02.297 "is_configured": true, 00:13:02.297 "data_offset": 2048, 00:13:02.297 "data_size": 63488 00:13:02.297 }, 00:13:02.297 { 00:13:02.297 "name": null, 00:13:02.297 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.297 "is_configured": false, 00:13:02.297 "data_offset": 2048, 00:13:02.297 "data_size": 63488 00:13:02.297 }, 00:13:02.297 { 00:13:02.297 "name": "BaseBdev3", 00:13:02.297 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:02.297 "is_configured": true, 00:13:02.297 "data_offset": 2048, 00:13:02.297 "data_size": 63488 00:13:02.297 }, 00:13:02.297 { 00:13:02.297 "name": "BaseBdev4", 00:13:02.297 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:02.297 "is_configured": true, 00:13:02.297 "data_offset": 2048, 00:13:02.297 "data_size": 63488 00:13:02.297 } 00:13:02.297 ] 00:13:02.297 }' 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:02.297 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:02.557 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:02.557 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:02.557 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.557 16:51:23 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.557 [2024-09-29 16:51:23.990139] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:02.557 [2024-09-29 16:51:24.053141] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:02.557 [2024-09-29 16:51:24.053261] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:02.557 [2024-09-29 16:51:24.053296] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:02.557 [2024-09-29 16:51:24.053319] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:02.557 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:02.558 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.558 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:02.558 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.558 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:02.558 "name": "raid_bdev1", 00:13:02.558 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:02.558 "strip_size_kb": 0, 00:13:02.558 "state": "online", 00:13:02.558 "raid_level": "raid1", 00:13:02.558 "superblock": true, 00:13:02.558 "num_base_bdevs": 4, 00:13:02.558 "num_base_bdevs_discovered": 2, 00:13:02.558 "num_base_bdevs_operational": 2, 00:13:02.558 "base_bdevs_list": [ 00:13:02.558 { 00:13:02.558 "name": null, 00:13:02.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.558 "is_configured": false, 00:13:02.558 "data_offset": 0, 00:13:02.558 "data_size": 63488 00:13:02.558 }, 00:13:02.558 { 00:13:02.558 "name": null, 00:13:02.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:02.558 "is_configured": false, 00:13:02.558 "data_offset": 2048, 00:13:02.558 "data_size": 63488 00:13:02.558 }, 00:13:02.558 { 00:13:02.558 "name": "BaseBdev3", 00:13:02.558 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:02.558 "is_configured": true, 00:13:02.558 "data_offset": 2048, 00:13:02.558 "data_size": 63488 00:13:02.558 }, 00:13:02.558 { 00:13:02.558 "name": "BaseBdev4", 00:13:02.558 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:02.558 "is_configured": true, 00:13:02.558 "data_offset": 2048, 00:13:02.558 "data_size": 63488 00:13:02.558 } 00:13:02.558 ] 00:13:02.558 }' 00:13:02.558 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:02.558 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.127 "name": "raid_bdev1", 00:13:03.127 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:03.127 "strip_size_kb": 0, 00:13:03.127 "state": "online", 00:13:03.127 "raid_level": "raid1", 00:13:03.127 "superblock": true, 00:13:03.127 "num_base_bdevs": 4, 00:13:03.127 "num_base_bdevs_discovered": 2, 00:13:03.127 "num_base_bdevs_operational": 2, 00:13:03.127 "base_bdevs_list": [ 00:13:03.127 { 00:13:03.127 "name": null, 00:13:03.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.127 "is_configured": false, 00:13:03.127 "data_offset": 0, 00:13:03.127 "data_size": 63488 00:13:03.127 }, 00:13:03.127 { 00:13:03.127 "name": null, 00:13:03.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.127 "is_configured": false, 00:13:03.127 "data_offset": 2048, 00:13:03.127 "data_size": 63488 00:13:03.127 }, 00:13:03.127 { 00:13:03.127 "name": "BaseBdev3", 00:13:03.127 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:03.127 "is_configured": true, 00:13:03.127 "data_offset": 2048, 00:13:03.127 "data_size": 63488 00:13:03.127 }, 00:13:03.127 { 00:13:03.127 "name": "BaseBdev4", 00:13:03.127 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:03.127 "is_configured": true, 00:13:03.127 "data_offset": 2048, 00:13:03.127 "data_size": 63488 00:13:03.127 } 00:13:03.127 ] 00:13:03.127 }' 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:03.127 [2024-09-29 16:51:24.628337] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:03.127 [2024-09-29 16:51:24.628465] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:03.127 [2024-09-29 16:51:24.628490] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:13:03.127 [2024-09-29 16:51:24.628500] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:03.127 [2024-09-29 16:51:24.628913] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:03.127 [2024-09-29 16:51:24.628939] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:03.127 [2024-09-29 16:51:24.629009] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:03.127 [2024-09-29 16:51:24.629024] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:03.127 [2024-09-29 16:51:24.629032] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:03.127 [2024-09-29 16:51:24.629050] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:03.127 BaseBdev1 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.127 16:51:24 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:04.067 "name": "raid_bdev1", 00:13:04.067 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:04.067 "strip_size_kb": 0, 00:13:04.067 "state": "online", 00:13:04.067 "raid_level": "raid1", 00:13:04.067 "superblock": true, 00:13:04.067 "num_base_bdevs": 4, 00:13:04.067 "num_base_bdevs_discovered": 2, 00:13:04.067 "num_base_bdevs_operational": 2, 00:13:04.067 "base_bdevs_list": [ 00:13:04.067 { 00:13:04.067 "name": null, 00:13:04.067 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.067 "is_configured": false, 00:13:04.067 "data_offset": 0, 00:13:04.067 "data_size": 63488 00:13:04.067 }, 00:13:04.067 { 00:13:04.067 "name": null, 00:13:04.067 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.067 "is_configured": false, 00:13:04.067 "data_offset": 2048, 00:13:04.067 "data_size": 63488 00:13:04.067 }, 00:13:04.067 { 00:13:04.067 "name": "BaseBdev3", 00:13:04.067 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:04.067 "is_configured": true, 00:13:04.067 "data_offset": 2048, 00:13:04.067 "data_size": 63488 00:13:04.067 }, 00:13:04.067 { 00:13:04.067 "name": "BaseBdev4", 00:13:04.067 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:04.067 "is_configured": true, 00:13:04.067 "data_offset": 2048, 00:13:04.067 "data_size": 63488 00:13:04.067 } 00:13:04.067 ] 00:13:04.067 }' 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:04.067 16:51:25 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:04.637 "name": "raid_bdev1", 00:13:04.637 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:04.637 "strip_size_kb": 0, 00:13:04.637 "state": "online", 00:13:04.637 "raid_level": "raid1", 00:13:04.637 "superblock": true, 00:13:04.637 "num_base_bdevs": 4, 00:13:04.637 "num_base_bdevs_discovered": 2, 00:13:04.637 "num_base_bdevs_operational": 2, 00:13:04.637 "base_bdevs_list": [ 00:13:04.637 { 00:13:04.637 "name": null, 00:13:04.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.637 "is_configured": false, 00:13:04.637 "data_offset": 0, 00:13:04.637 "data_size": 63488 00:13:04.637 }, 00:13:04.637 { 00:13:04.637 "name": null, 00:13:04.637 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:04.637 "is_configured": false, 00:13:04.637 "data_offset": 2048, 00:13:04.637 "data_size": 63488 00:13:04.637 }, 00:13:04.637 { 00:13:04.637 "name": "BaseBdev3", 00:13:04.637 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:04.637 "is_configured": true, 00:13:04.637 "data_offset": 2048, 00:13:04.637 "data_size": 63488 00:13:04.637 }, 00:13:04.637 { 00:13:04.637 "name": "BaseBdev4", 00:13:04.637 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:04.637 "is_configured": true, 00:13:04.637 "data_offset": 2048, 00:13:04.637 "data_size": 63488 00:13:04.637 } 00:13:04.637 ] 00:13:04.637 }' 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.637 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:04.637 [2024-09-29 16:51:26.242523] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:04.638 [2024-09-29 16:51:26.242750] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:04.638 [2024-09-29 16:51:26.242808] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:04.638 request: 00:13:04.638 { 00:13:04.638 "base_bdev": "BaseBdev1", 00:13:04.638 "raid_bdev": "raid_bdev1", 00:13:04.638 "method": "bdev_raid_add_base_bdev", 00:13:04.638 "req_id": 1 00:13:04.638 } 00:13:04.638 Got JSON-RPC error response 00:13:04.638 response: 00:13:04.638 { 00:13:04.638 "code": -22, 00:13:04.638 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:04.638 } 00:13:04.638 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:04.638 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:13:04.638 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:04.638 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:04.638 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:04.638 16:51:26 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:06.019 "name": "raid_bdev1", 00:13:06.019 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:06.019 "strip_size_kb": 0, 00:13:06.019 "state": "online", 00:13:06.019 "raid_level": "raid1", 00:13:06.019 "superblock": true, 00:13:06.019 "num_base_bdevs": 4, 00:13:06.019 "num_base_bdevs_discovered": 2, 00:13:06.019 "num_base_bdevs_operational": 2, 00:13:06.019 "base_bdevs_list": [ 00:13:06.019 { 00:13:06.019 "name": null, 00:13:06.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.019 "is_configured": false, 00:13:06.019 "data_offset": 0, 00:13:06.019 "data_size": 63488 00:13:06.019 }, 00:13:06.019 { 00:13:06.019 "name": null, 00:13:06.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.019 "is_configured": false, 00:13:06.019 "data_offset": 2048, 00:13:06.019 "data_size": 63488 00:13:06.019 }, 00:13:06.019 { 00:13:06.019 "name": "BaseBdev3", 00:13:06.019 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:06.019 "is_configured": true, 00:13:06.019 "data_offset": 2048, 00:13:06.019 "data_size": 63488 00:13:06.019 }, 00:13:06.019 { 00:13:06.019 "name": "BaseBdev4", 00:13:06.019 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:06.019 "is_configured": true, 00:13:06.019 "data_offset": 2048, 00:13:06.019 "data_size": 63488 00:13:06.019 } 00:13:06.019 ] 00:13:06.019 }' 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:06.019 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.280 "name": "raid_bdev1", 00:13:06.280 "uuid": "140ad76e-a802-42f1-980f-2dc7a313ff71", 00:13:06.280 "strip_size_kb": 0, 00:13:06.280 "state": "online", 00:13:06.280 "raid_level": "raid1", 00:13:06.280 "superblock": true, 00:13:06.280 "num_base_bdevs": 4, 00:13:06.280 "num_base_bdevs_discovered": 2, 00:13:06.280 "num_base_bdevs_operational": 2, 00:13:06.280 "base_bdevs_list": [ 00:13:06.280 { 00:13:06.280 "name": null, 00:13:06.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.280 "is_configured": false, 00:13:06.280 "data_offset": 0, 00:13:06.280 "data_size": 63488 00:13:06.280 }, 00:13:06.280 { 00:13:06.280 "name": null, 00:13:06.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.280 "is_configured": false, 00:13:06.280 "data_offset": 2048, 00:13:06.280 "data_size": 63488 00:13:06.280 }, 00:13:06.280 { 00:13:06.280 "name": "BaseBdev3", 00:13:06.280 "uuid": "5b516f2a-cc10-524c-aa42-21e55b3ae690", 00:13:06.280 "is_configured": true, 00:13:06.280 "data_offset": 2048, 00:13:06.280 "data_size": 63488 00:13:06.280 }, 00:13:06.280 { 00:13:06.280 "name": "BaseBdev4", 00:13:06.280 "uuid": "fd0f4526-caa0-561d-b539-c5d620c6b844", 00:13:06.280 "is_configured": true, 00:13:06.280 "data_offset": 2048, 00:13:06.280 "data_size": 63488 00:13:06.280 } 00:13:06.280 ] 00:13:06.280 }' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 89383 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 89383 ']' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 89383 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89383 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:06.280 killing process with pid 89383 00:13:06.280 Received shutdown signal, test time was about 17.793759 seconds 00:13:06.280 00:13:06.280 Latency(us) 00:13:06.280 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:06.280 =================================================================================================================== 00:13:06.280 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89383' 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 89383 00:13:06.280 [2024-09-29 16:51:27.894455] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:06.280 16:51:27 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 89383 00:13:06.280 [2024-09-29 16:51:27.894573] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:06.280 [2024-09-29 16:51:27.894648] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:06.280 [2024-09-29 16:51:27.894674] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:06.280 [2024-09-29 16:51:27.940467] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:06.540 16:51:28 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:06.540 00:13:06.540 real 0m19.812s 00:13:06.540 user 0m26.408s 00:13:06.540 sys 0m2.575s 00:13:06.540 16:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:06.540 16:51:28 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:06.540 ************************************ 00:13:06.540 END TEST raid_rebuild_test_sb_io 00:13:06.540 ************************************ 00:13:06.801 16:51:28 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:06.801 16:51:28 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:13:06.801 16:51:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:13:06.801 16:51:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:06.801 16:51:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:06.801 ************************************ 00:13:06.801 START TEST raid5f_state_function_test 00:13:06.801 ************************************ 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 false 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:06.801 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90094 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90094' 00:13:06.802 Process raid pid: 90094 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90094 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 90094 ']' 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:06.802 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:06.802 16:51:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:06.802 [2024-09-29 16:51:28.352703] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:13:06.802 [2024-09-29 16:51:28.352920] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:07.062 [2024-09-29 16:51:28.500189] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:07.062 [2024-09-29 16:51:28.547104] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:07.062 [2024-09-29 16:51:28.589685] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:07.062 [2024-09-29 16:51:28.589808] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.631 [2024-09-29 16:51:29.183411] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:07.631 [2024-09-29 16:51:29.183455] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:07.631 [2024-09-29 16:51:29.183467] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:07.631 [2024-09-29 16:51:29.183476] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:07.631 [2024-09-29 16:51:29.183482] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:07.631 [2024-09-29 16:51:29.183494] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:07.631 "name": "Existed_Raid", 00:13:07.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.631 "strip_size_kb": 64, 00:13:07.631 "state": "configuring", 00:13:07.631 "raid_level": "raid5f", 00:13:07.631 "superblock": false, 00:13:07.631 "num_base_bdevs": 3, 00:13:07.631 "num_base_bdevs_discovered": 0, 00:13:07.631 "num_base_bdevs_operational": 3, 00:13:07.631 "base_bdevs_list": [ 00:13:07.631 { 00:13:07.631 "name": "BaseBdev1", 00:13:07.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.631 "is_configured": false, 00:13:07.631 "data_offset": 0, 00:13:07.631 "data_size": 0 00:13:07.631 }, 00:13:07.631 { 00:13:07.631 "name": "BaseBdev2", 00:13:07.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.631 "is_configured": false, 00:13:07.631 "data_offset": 0, 00:13:07.631 "data_size": 0 00:13:07.631 }, 00:13:07.631 { 00:13:07.631 "name": "BaseBdev3", 00:13:07.631 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.631 "is_configured": false, 00:13:07.631 "data_offset": 0, 00:13:07.631 "data_size": 0 00:13:07.631 } 00:13:07.631 ] 00:13:07.631 }' 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:07.631 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.201 [2024-09-29 16:51:29.638589] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:08.201 [2024-09-29 16:51:29.638662] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.201 [2024-09-29 16:51:29.650587] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:08.201 [2024-09-29 16:51:29.650657] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:08.201 [2024-09-29 16:51:29.650698] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:08.201 [2024-09-29 16:51:29.650720] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:08.201 [2024-09-29 16:51:29.650745] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:08.201 [2024-09-29 16:51:29.650766] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.201 [2024-09-29 16:51:29.671718] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:08.201 BaseBdev1 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.201 [ 00:13:08.201 { 00:13:08.201 "name": "BaseBdev1", 00:13:08.201 "aliases": [ 00:13:08.201 "a497d33f-61de-459c-ac8e-0cb93b6031d9" 00:13:08.201 ], 00:13:08.201 "product_name": "Malloc disk", 00:13:08.201 "block_size": 512, 00:13:08.201 "num_blocks": 65536, 00:13:08.201 "uuid": "a497d33f-61de-459c-ac8e-0cb93b6031d9", 00:13:08.201 "assigned_rate_limits": { 00:13:08.201 "rw_ios_per_sec": 0, 00:13:08.201 "rw_mbytes_per_sec": 0, 00:13:08.201 "r_mbytes_per_sec": 0, 00:13:08.201 "w_mbytes_per_sec": 0 00:13:08.201 }, 00:13:08.201 "claimed": true, 00:13:08.201 "claim_type": "exclusive_write", 00:13:08.201 "zoned": false, 00:13:08.201 "supported_io_types": { 00:13:08.201 "read": true, 00:13:08.201 "write": true, 00:13:08.201 "unmap": true, 00:13:08.201 "flush": true, 00:13:08.201 "reset": true, 00:13:08.201 "nvme_admin": false, 00:13:08.201 "nvme_io": false, 00:13:08.201 "nvme_io_md": false, 00:13:08.201 "write_zeroes": true, 00:13:08.201 "zcopy": true, 00:13:08.201 "get_zone_info": false, 00:13:08.201 "zone_management": false, 00:13:08.201 "zone_append": false, 00:13:08.201 "compare": false, 00:13:08.201 "compare_and_write": false, 00:13:08.201 "abort": true, 00:13:08.201 "seek_hole": false, 00:13:08.201 "seek_data": false, 00:13:08.201 "copy": true, 00:13:08.201 "nvme_iov_md": false 00:13:08.201 }, 00:13:08.201 "memory_domains": [ 00:13:08.201 { 00:13:08.201 "dma_device_id": "system", 00:13:08.201 "dma_device_type": 1 00:13:08.201 }, 00:13:08.201 { 00:13:08.201 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:08.201 "dma_device_type": 2 00:13:08.201 } 00:13:08.201 ], 00:13:08.201 "driver_specific": {} 00:13:08.201 } 00:13:08.201 ] 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:08.201 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:08.202 "name": "Existed_Raid", 00:13:08.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.202 "strip_size_kb": 64, 00:13:08.202 "state": "configuring", 00:13:08.202 "raid_level": "raid5f", 00:13:08.202 "superblock": false, 00:13:08.202 "num_base_bdevs": 3, 00:13:08.202 "num_base_bdevs_discovered": 1, 00:13:08.202 "num_base_bdevs_operational": 3, 00:13:08.202 "base_bdevs_list": [ 00:13:08.202 { 00:13:08.202 "name": "BaseBdev1", 00:13:08.202 "uuid": "a497d33f-61de-459c-ac8e-0cb93b6031d9", 00:13:08.202 "is_configured": true, 00:13:08.202 "data_offset": 0, 00:13:08.202 "data_size": 65536 00:13:08.202 }, 00:13:08.202 { 00:13:08.202 "name": "BaseBdev2", 00:13:08.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.202 "is_configured": false, 00:13:08.202 "data_offset": 0, 00:13:08.202 "data_size": 0 00:13:08.202 }, 00:13:08.202 { 00:13:08.202 "name": "BaseBdev3", 00:13:08.202 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.202 "is_configured": false, 00:13:08.202 "data_offset": 0, 00:13:08.202 "data_size": 0 00:13:08.202 } 00:13:08.202 ] 00:13:08.202 }' 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:08.202 16:51:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.771 [2024-09-29 16:51:30.154906] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:08.771 [2024-09-29 16:51:30.155013] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.771 [2024-09-29 16:51:30.166941] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:08.771 [2024-09-29 16:51:30.168816] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:08.771 [2024-09-29 16:51:30.168886] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:08.771 [2024-09-29 16:51:30.168912] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:08.771 [2024-09-29 16:51:30.168935] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:08.771 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:08.772 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.772 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:08.772 "name": "Existed_Raid", 00:13:08.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.772 "strip_size_kb": 64, 00:13:08.772 "state": "configuring", 00:13:08.772 "raid_level": "raid5f", 00:13:08.772 "superblock": false, 00:13:08.772 "num_base_bdevs": 3, 00:13:08.772 "num_base_bdevs_discovered": 1, 00:13:08.772 "num_base_bdevs_operational": 3, 00:13:08.772 "base_bdevs_list": [ 00:13:08.772 { 00:13:08.772 "name": "BaseBdev1", 00:13:08.772 "uuid": "a497d33f-61de-459c-ac8e-0cb93b6031d9", 00:13:08.772 "is_configured": true, 00:13:08.772 "data_offset": 0, 00:13:08.772 "data_size": 65536 00:13:08.772 }, 00:13:08.772 { 00:13:08.772 "name": "BaseBdev2", 00:13:08.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.772 "is_configured": false, 00:13:08.772 "data_offset": 0, 00:13:08.772 "data_size": 0 00:13:08.772 }, 00:13:08.772 { 00:13:08.772 "name": "BaseBdev3", 00:13:08.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:08.772 "is_configured": false, 00:13:08.772 "data_offset": 0, 00:13:08.772 "data_size": 0 00:13:08.772 } 00:13:08.772 ] 00:13:08.772 }' 00:13:08.772 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:08.772 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.032 [2024-09-29 16:51:30.676842] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:09.032 BaseBdev2 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.032 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.032 [ 00:13:09.032 { 00:13:09.032 "name": "BaseBdev2", 00:13:09.032 "aliases": [ 00:13:09.032 "46170411-efad-42c5-8f82-8f98e3402f70" 00:13:09.032 ], 00:13:09.032 "product_name": "Malloc disk", 00:13:09.032 "block_size": 512, 00:13:09.032 "num_blocks": 65536, 00:13:09.032 "uuid": "46170411-efad-42c5-8f82-8f98e3402f70", 00:13:09.032 "assigned_rate_limits": { 00:13:09.032 "rw_ios_per_sec": 0, 00:13:09.292 "rw_mbytes_per_sec": 0, 00:13:09.292 "r_mbytes_per_sec": 0, 00:13:09.292 "w_mbytes_per_sec": 0 00:13:09.292 }, 00:13:09.292 "claimed": true, 00:13:09.292 "claim_type": "exclusive_write", 00:13:09.292 "zoned": false, 00:13:09.292 "supported_io_types": { 00:13:09.292 "read": true, 00:13:09.292 "write": true, 00:13:09.292 "unmap": true, 00:13:09.292 "flush": true, 00:13:09.292 "reset": true, 00:13:09.292 "nvme_admin": false, 00:13:09.292 "nvme_io": false, 00:13:09.292 "nvme_io_md": false, 00:13:09.292 "write_zeroes": true, 00:13:09.292 "zcopy": true, 00:13:09.292 "get_zone_info": false, 00:13:09.292 "zone_management": false, 00:13:09.292 "zone_append": false, 00:13:09.292 "compare": false, 00:13:09.292 "compare_and_write": false, 00:13:09.292 "abort": true, 00:13:09.292 "seek_hole": false, 00:13:09.292 "seek_data": false, 00:13:09.292 "copy": true, 00:13:09.292 "nvme_iov_md": false 00:13:09.292 }, 00:13:09.292 "memory_domains": [ 00:13:09.292 { 00:13:09.292 "dma_device_id": "system", 00:13:09.292 "dma_device_type": 1 00:13:09.292 }, 00:13:09.292 { 00:13:09.292 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:09.292 "dma_device_type": 2 00:13:09.292 } 00:13:09.292 ], 00:13:09.292 "driver_specific": {} 00:13:09.292 } 00:13:09.292 ] 00:13:09.292 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.292 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:09.292 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:09.293 "name": "Existed_Raid", 00:13:09.293 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.293 "strip_size_kb": 64, 00:13:09.293 "state": "configuring", 00:13:09.293 "raid_level": "raid5f", 00:13:09.293 "superblock": false, 00:13:09.293 "num_base_bdevs": 3, 00:13:09.293 "num_base_bdevs_discovered": 2, 00:13:09.293 "num_base_bdevs_operational": 3, 00:13:09.293 "base_bdevs_list": [ 00:13:09.293 { 00:13:09.293 "name": "BaseBdev1", 00:13:09.293 "uuid": "a497d33f-61de-459c-ac8e-0cb93b6031d9", 00:13:09.293 "is_configured": true, 00:13:09.293 "data_offset": 0, 00:13:09.293 "data_size": 65536 00:13:09.293 }, 00:13:09.293 { 00:13:09.293 "name": "BaseBdev2", 00:13:09.293 "uuid": "46170411-efad-42c5-8f82-8f98e3402f70", 00:13:09.293 "is_configured": true, 00:13:09.293 "data_offset": 0, 00:13:09.293 "data_size": 65536 00:13:09.293 }, 00:13:09.293 { 00:13:09.293 "name": "BaseBdev3", 00:13:09.293 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.293 "is_configured": false, 00:13:09.293 "data_offset": 0, 00:13:09.293 "data_size": 0 00:13:09.293 } 00:13:09.293 ] 00:13:09.293 }' 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:09.293 16:51:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.553 [2024-09-29 16:51:31.175019] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:09.553 [2024-09-29 16:51:31.175123] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:09.553 [2024-09-29 16:51:31.175152] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:09.553 [2024-09-29 16:51:31.175503] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:09.553 [2024-09-29 16:51:31.176042] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:09.553 [2024-09-29 16:51:31.176095] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:09.553 [2024-09-29 16:51:31.176369] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:09.553 BaseBdev3 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.553 [ 00:13:09.553 { 00:13:09.553 "name": "BaseBdev3", 00:13:09.553 "aliases": [ 00:13:09.553 "159c1d59-fbce-4a5e-98eb-0a749271867b" 00:13:09.553 ], 00:13:09.553 "product_name": "Malloc disk", 00:13:09.553 "block_size": 512, 00:13:09.553 "num_blocks": 65536, 00:13:09.553 "uuid": "159c1d59-fbce-4a5e-98eb-0a749271867b", 00:13:09.553 "assigned_rate_limits": { 00:13:09.553 "rw_ios_per_sec": 0, 00:13:09.553 "rw_mbytes_per_sec": 0, 00:13:09.553 "r_mbytes_per_sec": 0, 00:13:09.553 "w_mbytes_per_sec": 0 00:13:09.553 }, 00:13:09.553 "claimed": true, 00:13:09.553 "claim_type": "exclusive_write", 00:13:09.553 "zoned": false, 00:13:09.553 "supported_io_types": { 00:13:09.553 "read": true, 00:13:09.553 "write": true, 00:13:09.553 "unmap": true, 00:13:09.553 "flush": true, 00:13:09.553 "reset": true, 00:13:09.553 "nvme_admin": false, 00:13:09.553 "nvme_io": false, 00:13:09.553 "nvme_io_md": false, 00:13:09.553 "write_zeroes": true, 00:13:09.553 "zcopy": true, 00:13:09.553 "get_zone_info": false, 00:13:09.553 "zone_management": false, 00:13:09.553 "zone_append": false, 00:13:09.553 "compare": false, 00:13:09.553 "compare_and_write": false, 00:13:09.553 "abort": true, 00:13:09.553 "seek_hole": false, 00:13:09.553 "seek_data": false, 00:13:09.553 "copy": true, 00:13:09.553 "nvme_iov_md": false 00:13:09.553 }, 00:13:09.553 "memory_domains": [ 00:13:09.553 { 00:13:09.553 "dma_device_id": "system", 00:13:09.553 "dma_device_type": 1 00:13:09.553 }, 00:13:09.553 { 00:13:09.553 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:09.553 "dma_device_type": 2 00:13:09.553 } 00:13:09.553 ], 00:13:09.553 "driver_specific": {} 00:13:09.553 } 00:13:09.553 ] 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:09.553 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:09.813 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.813 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:09.813 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.813 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:09.813 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.813 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:09.813 "name": "Existed_Raid", 00:13:09.813 "uuid": "ead21349-12e2-428a-8967-b94b49d43b0e", 00:13:09.813 "strip_size_kb": 64, 00:13:09.813 "state": "online", 00:13:09.813 "raid_level": "raid5f", 00:13:09.813 "superblock": false, 00:13:09.813 "num_base_bdevs": 3, 00:13:09.813 "num_base_bdevs_discovered": 3, 00:13:09.813 "num_base_bdevs_operational": 3, 00:13:09.813 "base_bdevs_list": [ 00:13:09.813 { 00:13:09.813 "name": "BaseBdev1", 00:13:09.813 "uuid": "a497d33f-61de-459c-ac8e-0cb93b6031d9", 00:13:09.813 "is_configured": true, 00:13:09.813 "data_offset": 0, 00:13:09.813 "data_size": 65536 00:13:09.813 }, 00:13:09.813 { 00:13:09.813 "name": "BaseBdev2", 00:13:09.813 "uuid": "46170411-efad-42c5-8f82-8f98e3402f70", 00:13:09.813 "is_configured": true, 00:13:09.814 "data_offset": 0, 00:13:09.814 "data_size": 65536 00:13:09.814 }, 00:13:09.814 { 00:13:09.814 "name": "BaseBdev3", 00:13:09.814 "uuid": "159c1d59-fbce-4a5e-98eb-0a749271867b", 00:13:09.814 "is_configured": true, 00:13:09.814 "data_offset": 0, 00:13:09.814 "data_size": 65536 00:13:09.814 } 00:13:09.814 ] 00:13:09.814 }' 00:13:09.814 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:09.814 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.074 [2024-09-29 16:51:31.666368] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:10.074 "name": "Existed_Raid", 00:13:10.074 "aliases": [ 00:13:10.074 "ead21349-12e2-428a-8967-b94b49d43b0e" 00:13:10.074 ], 00:13:10.074 "product_name": "Raid Volume", 00:13:10.074 "block_size": 512, 00:13:10.074 "num_blocks": 131072, 00:13:10.074 "uuid": "ead21349-12e2-428a-8967-b94b49d43b0e", 00:13:10.074 "assigned_rate_limits": { 00:13:10.074 "rw_ios_per_sec": 0, 00:13:10.074 "rw_mbytes_per_sec": 0, 00:13:10.074 "r_mbytes_per_sec": 0, 00:13:10.074 "w_mbytes_per_sec": 0 00:13:10.074 }, 00:13:10.074 "claimed": false, 00:13:10.074 "zoned": false, 00:13:10.074 "supported_io_types": { 00:13:10.074 "read": true, 00:13:10.074 "write": true, 00:13:10.074 "unmap": false, 00:13:10.074 "flush": false, 00:13:10.074 "reset": true, 00:13:10.074 "nvme_admin": false, 00:13:10.074 "nvme_io": false, 00:13:10.074 "nvme_io_md": false, 00:13:10.074 "write_zeroes": true, 00:13:10.074 "zcopy": false, 00:13:10.074 "get_zone_info": false, 00:13:10.074 "zone_management": false, 00:13:10.074 "zone_append": false, 00:13:10.074 "compare": false, 00:13:10.074 "compare_and_write": false, 00:13:10.074 "abort": false, 00:13:10.074 "seek_hole": false, 00:13:10.074 "seek_data": false, 00:13:10.074 "copy": false, 00:13:10.074 "nvme_iov_md": false 00:13:10.074 }, 00:13:10.074 "driver_specific": { 00:13:10.074 "raid": { 00:13:10.074 "uuid": "ead21349-12e2-428a-8967-b94b49d43b0e", 00:13:10.074 "strip_size_kb": 64, 00:13:10.074 "state": "online", 00:13:10.074 "raid_level": "raid5f", 00:13:10.074 "superblock": false, 00:13:10.074 "num_base_bdevs": 3, 00:13:10.074 "num_base_bdevs_discovered": 3, 00:13:10.074 "num_base_bdevs_operational": 3, 00:13:10.074 "base_bdevs_list": [ 00:13:10.074 { 00:13:10.074 "name": "BaseBdev1", 00:13:10.074 "uuid": "a497d33f-61de-459c-ac8e-0cb93b6031d9", 00:13:10.074 "is_configured": true, 00:13:10.074 "data_offset": 0, 00:13:10.074 "data_size": 65536 00:13:10.074 }, 00:13:10.074 { 00:13:10.074 "name": "BaseBdev2", 00:13:10.074 "uuid": "46170411-efad-42c5-8f82-8f98e3402f70", 00:13:10.074 "is_configured": true, 00:13:10.074 "data_offset": 0, 00:13:10.074 "data_size": 65536 00:13:10.074 }, 00:13:10.074 { 00:13:10.074 "name": "BaseBdev3", 00:13:10.074 "uuid": "159c1d59-fbce-4a5e-98eb-0a749271867b", 00:13:10.074 "is_configured": true, 00:13:10.074 "data_offset": 0, 00:13:10.074 "data_size": 65536 00:13:10.074 } 00:13:10.074 ] 00:13:10.074 } 00:13:10.074 } 00:13:10.074 }' 00:13:10.074 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:10.335 BaseBdev2 00:13:10.335 BaseBdev3' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.335 [2024-09-29 16:51:31.965762] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:10.335 16:51:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.595 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:10.595 "name": "Existed_Raid", 00:13:10.595 "uuid": "ead21349-12e2-428a-8967-b94b49d43b0e", 00:13:10.595 "strip_size_kb": 64, 00:13:10.595 "state": "online", 00:13:10.595 "raid_level": "raid5f", 00:13:10.595 "superblock": false, 00:13:10.595 "num_base_bdevs": 3, 00:13:10.595 "num_base_bdevs_discovered": 2, 00:13:10.595 "num_base_bdevs_operational": 2, 00:13:10.595 "base_bdevs_list": [ 00:13:10.595 { 00:13:10.595 "name": null, 00:13:10.595 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.595 "is_configured": false, 00:13:10.595 "data_offset": 0, 00:13:10.595 "data_size": 65536 00:13:10.595 }, 00:13:10.595 { 00:13:10.595 "name": "BaseBdev2", 00:13:10.595 "uuid": "46170411-efad-42c5-8f82-8f98e3402f70", 00:13:10.595 "is_configured": true, 00:13:10.595 "data_offset": 0, 00:13:10.595 "data_size": 65536 00:13:10.595 }, 00:13:10.595 { 00:13:10.595 "name": "BaseBdev3", 00:13:10.595 "uuid": "159c1d59-fbce-4a5e-98eb-0a749271867b", 00:13:10.595 "is_configured": true, 00:13:10.595 "data_offset": 0, 00:13:10.595 "data_size": 65536 00:13:10.595 } 00:13:10.595 ] 00:13:10.595 }' 00:13:10.595 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:10.595 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:10.855 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:10.856 [2024-09-29 16:51:32.504287] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:10.856 [2024-09-29 16:51:32.504452] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:10.856 [2024-09-29 16:51:32.515774] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:10.856 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.116 [2024-09-29 16:51:32.571686] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:11.116 [2024-09-29 16:51:32.571814] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.116 BaseBdev2 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.116 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.116 [ 00:13:11.116 { 00:13:11.116 "name": "BaseBdev2", 00:13:11.116 "aliases": [ 00:13:11.116 "4f054a1d-fba6-4552-bfa9-9319854bd7a5" 00:13:11.116 ], 00:13:11.116 "product_name": "Malloc disk", 00:13:11.117 "block_size": 512, 00:13:11.117 "num_blocks": 65536, 00:13:11.117 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:11.117 "assigned_rate_limits": { 00:13:11.117 "rw_ios_per_sec": 0, 00:13:11.117 "rw_mbytes_per_sec": 0, 00:13:11.117 "r_mbytes_per_sec": 0, 00:13:11.117 "w_mbytes_per_sec": 0 00:13:11.117 }, 00:13:11.117 "claimed": false, 00:13:11.117 "zoned": false, 00:13:11.117 "supported_io_types": { 00:13:11.117 "read": true, 00:13:11.117 "write": true, 00:13:11.117 "unmap": true, 00:13:11.117 "flush": true, 00:13:11.117 "reset": true, 00:13:11.117 "nvme_admin": false, 00:13:11.117 "nvme_io": false, 00:13:11.117 "nvme_io_md": false, 00:13:11.117 "write_zeroes": true, 00:13:11.117 "zcopy": true, 00:13:11.117 "get_zone_info": false, 00:13:11.117 "zone_management": false, 00:13:11.117 "zone_append": false, 00:13:11.117 "compare": false, 00:13:11.117 "compare_and_write": false, 00:13:11.117 "abort": true, 00:13:11.117 "seek_hole": false, 00:13:11.117 "seek_data": false, 00:13:11.117 "copy": true, 00:13:11.117 "nvme_iov_md": false 00:13:11.117 }, 00:13:11.117 "memory_domains": [ 00:13:11.117 { 00:13:11.117 "dma_device_id": "system", 00:13:11.117 "dma_device_type": 1 00:13:11.117 }, 00:13:11.117 { 00:13:11.117 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:11.117 "dma_device_type": 2 00:13:11.117 } 00:13:11.117 ], 00:13:11.117 "driver_specific": {} 00:13:11.117 } 00:13:11.117 ] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.117 BaseBdev3 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.117 [ 00:13:11.117 { 00:13:11.117 "name": "BaseBdev3", 00:13:11.117 "aliases": [ 00:13:11.117 "26dfe549-8423-42f6-b123-bb7d750e2524" 00:13:11.117 ], 00:13:11.117 "product_name": "Malloc disk", 00:13:11.117 "block_size": 512, 00:13:11.117 "num_blocks": 65536, 00:13:11.117 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:11.117 "assigned_rate_limits": { 00:13:11.117 "rw_ios_per_sec": 0, 00:13:11.117 "rw_mbytes_per_sec": 0, 00:13:11.117 "r_mbytes_per_sec": 0, 00:13:11.117 "w_mbytes_per_sec": 0 00:13:11.117 }, 00:13:11.117 "claimed": false, 00:13:11.117 "zoned": false, 00:13:11.117 "supported_io_types": { 00:13:11.117 "read": true, 00:13:11.117 "write": true, 00:13:11.117 "unmap": true, 00:13:11.117 "flush": true, 00:13:11.117 "reset": true, 00:13:11.117 "nvme_admin": false, 00:13:11.117 "nvme_io": false, 00:13:11.117 "nvme_io_md": false, 00:13:11.117 "write_zeroes": true, 00:13:11.117 "zcopy": true, 00:13:11.117 "get_zone_info": false, 00:13:11.117 "zone_management": false, 00:13:11.117 "zone_append": false, 00:13:11.117 "compare": false, 00:13:11.117 "compare_and_write": false, 00:13:11.117 "abort": true, 00:13:11.117 "seek_hole": false, 00:13:11.117 "seek_data": false, 00:13:11.117 "copy": true, 00:13:11.117 "nvme_iov_md": false 00:13:11.117 }, 00:13:11.117 "memory_domains": [ 00:13:11.117 { 00:13:11.117 "dma_device_id": "system", 00:13:11.117 "dma_device_type": 1 00:13:11.117 }, 00:13:11.117 { 00:13:11.117 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:11.117 "dma_device_type": 2 00:13:11.117 } 00:13:11.117 ], 00:13:11.117 "driver_specific": {} 00:13:11.117 } 00:13:11.117 ] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.117 [2024-09-29 16:51:32.746067] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:11.117 [2024-09-29 16:51:32.746199] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:11.117 [2024-09-29 16:51:32.746239] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:11.117 [2024-09-29 16:51:32.748054] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.117 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.377 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.377 "name": "Existed_Raid", 00:13:11.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.378 "strip_size_kb": 64, 00:13:11.378 "state": "configuring", 00:13:11.378 "raid_level": "raid5f", 00:13:11.378 "superblock": false, 00:13:11.378 "num_base_bdevs": 3, 00:13:11.378 "num_base_bdevs_discovered": 2, 00:13:11.378 "num_base_bdevs_operational": 3, 00:13:11.378 "base_bdevs_list": [ 00:13:11.378 { 00:13:11.378 "name": "BaseBdev1", 00:13:11.378 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.378 "is_configured": false, 00:13:11.378 "data_offset": 0, 00:13:11.378 "data_size": 0 00:13:11.378 }, 00:13:11.378 { 00:13:11.378 "name": "BaseBdev2", 00:13:11.378 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:11.378 "is_configured": true, 00:13:11.378 "data_offset": 0, 00:13:11.378 "data_size": 65536 00:13:11.378 }, 00:13:11.378 { 00:13:11.378 "name": "BaseBdev3", 00:13:11.378 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:11.378 "is_configured": true, 00:13:11.378 "data_offset": 0, 00:13:11.378 "data_size": 65536 00:13:11.378 } 00:13:11.378 ] 00:13:11.378 }' 00:13:11.378 16:51:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.378 16:51:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.638 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:11.638 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.638 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.638 [2024-09-29 16:51:33.229189] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:11.638 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.638 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:11.638 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.639 "name": "Existed_Raid", 00:13:11.639 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.639 "strip_size_kb": 64, 00:13:11.639 "state": "configuring", 00:13:11.639 "raid_level": "raid5f", 00:13:11.639 "superblock": false, 00:13:11.639 "num_base_bdevs": 3, 00:13:11.639 "num_base_bdevs_discovered": 1, 00:13:11.639 "num_base_bdevs_operational": 3, 00:13:11.639 "base_bdevs_list": [ 00:13:11.639 { 00:13:11.639 "name": "BaseBdev1", 00:13:11.639 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.639 "is_configured": false, 00:13:11.639 "data_offset": 0, 00:13:11.639 "data_size": 0 00:13:11.639 }, 00:13:11.639 { 00:13:11.639 "name": null, 00:13:11.639 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:11.639 "is_configured": false, 00:13:11.639 "data_offset": 0, 00:13:11.639 "data_size": 65536 00:13:11.639 }, 00:13:11.639 { 00:13:11.639 "name": "BaseBdev3", 00:13:11.639 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:11.639 "is_configured": true, 00:13:11.639 "data_offset": 0, 00:13:11.639 "data_size": 65536 00:13:11.639 } 00:13:11.639 ] 00:13:11.639 }' 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.639 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.265 [2024-09-29 16:51:33.771430] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:12.265 BaseBdev1 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.265 [ 00:13:12.265 { 00:13:12.265 "name": "BaseBdev1", 00:13:12.265 "aliases": [ 00:13:12.265 "331f9cfb-d794-4513-b8eb-43940f90e306" 00:13:12.265 ], 00:13:12.265 "product_name": "Malloc disk", 00:13:12.265 "block_size": 512, 00:13:12.265 "num_blocks": 65536, 00:13:12.265 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:12.265 "assigned_rate_limits": { 00:13:12.265 "rw_ios_per_sec": 0, 00:13:12.265 "rw_mbytes_per_sec": 0, 00:13:12.265 "r_mbytes_per_sec": 0, 00:13:12.265 "w_mbytes_per_sec": 0 00:13:12.265 }, 00:13:12.265 "claimed": true, 00:13:12.265 "claim_type": "exclusive_write", 00:13:12.265 "zoned": false, 00:13:12.265 "supported_io_types": { 00:13:12.265 "read": true, 00:13:12.265 "write": true, 00:13:12.265 "unmap": true, 00:13:12.265 "flush": true, 00:13:12.265 "reset": true, 00:13:12.265 "nvme_admin": false, 00:13:12.265 "nvme_io": false, 00:13:12.265 "nvme_io_md": false, 00:13:12.265 "write_zeroes": true, 00:13:12.265 "zcopy": true, 00:13:12.265 "get_zone_info": false, 00:13:12.265 "zone_management": false, 00:13:12.265 "zone_append": false, 00:13:12.265 "compare": false, 00:13:12.265 "compare_and_write": false, 00:13:12.265 "abort": true, 00:13:12.265 "seek_hole": false, 00:13:12.265 "seek_data": false, 00:13:12.265 "copy": true, 00:13:12.265 "nvme_iov_md": false 00:13:12.265 }, 00:13:12.265 "memory_domains": [ 00:13:12.265 { 00:13:12.265 "dma_device_id": "system", 00:13:12.265 "dma_device_type": 1 00:13:12.265 }, 00:13:12.265 { 00:13:12.265 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:12.265 "dma_device_type": 2 00:13:12.265 } 00:13:12.265 ], 00:13:12.265 "driver_specific": {} 00:13:12.265 } 00:13:12.265 ] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.265 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.265 "name": "Existed_Raid", 00:13:12.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.265 "strip_size_kb": 64, 00:13:12.265 "state": "configuring", 00:13:12.265 "raid_level": "raid5f", 00:13:12.265 "superblock": false, 00:13:12.265 "num_base_bdevs": 3, 00:13:12.265 "num_base_bdevs_discovered": 2, 00:13:12.265 "num_base_bdevs_operational": 3, 00:13:12.265 "base_bdevs_list": [ 00:13:12.265 { 00:13:12.265 "name": "BaseBdev1", 00:13:12.265 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:12.265 "is_configured": true, 00:13:12.265 "data_offset": 0, 00:13:12.265 "data_size": 65536 00:13:12.265 }, 00:13:12.265 { 00:13:12.265 "name": null, 00:13:12.265 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:12.265 "is_configured": false, 00:13:12.265 "data_offset": 0, 00:13:12.265 "data_size": 65536 00:13:12.265 }, 00:13:12.265 { 00:13:12.265 "name": "BaseBdev3", 00:13:12.265 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:12.265 "is_configured": true, 00:13:12.265 "data_offset": 0, 00:13:12.266 "data_size": 65536 00:13:12.266 } 00:13:12.266 ] 00:13:12.266 }' 00:13:12.266 16:51:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.266 16:51:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.882 [2024-09-29 16:51:34.270994] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:12.882 "name": "Existed_Raid", 00:13:12.882 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:12.882 "strip_size_kb": 64, 00:13:12.882 "state": "configuring", 00:13:12.882 "raid_level": "raid5f", 00:13:12.882 "superblock": false, 00:13:12.882 "num_base_bdevs": 3, 00:13:12.882 "num_base_bdevs_discovered": 1, 00:13:12.882 "num_base_bdevs_operational": 3, 00:13:12.882 "base_bdevs_list": [ 00:13:12.882 { 00:13:12.882 "name": "BaseBdev1", 00:13:12.882 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:12.882 "is_configured": true, 00:13:12.882 "data_offset": 0, 00:13:12.882 "data_size": 65536 00:13:12.882 }, 00:13:12.882 { 00:13:12.882 "name": null, 00:13:12.882 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:12.882 "is_configured": false, 00:13:12.882 "data_offset": 0, 00:13:12.882 "data_size": 65536 00:13:12.882 }, 00:13:12.882 { 00:13:12.882 "name": null, 00:13:12.882 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:12.882 "is_configured": false, 00:13:12.882 "data_offset": 0, 00:13:12.882 "data_size": 65536 00:13:12.882 } 00:13:12.882 ] 00:13:12.882 }' 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:12.882 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.141 [2024-09-29 16:51:34.770156] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.141 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.399 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.399 "name": "Existed_Raid", 00:13:13.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.399 "strip_size_kb": 64, 00:13:13.399 "state": "configuring", 00:13:13.399 "raid_level": "raid5f", 00:13:13.399 "superblock": false, 00:13:13.399 "num_base_bdevs": 3, 00:13:13.399 "num_base_bdevs_discovered": 2, 00:13:13.399 "num_base_bdevs_operational": 3, 00:13:13.399 "base_bdevs_list": [ 00:13:13.399 { 00:13:13.399 "name": "BaseBdev1", 00:13:13.399 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:13.399 "is_configured": true, 00:13:13.399 "data_offset": 0, 00:13:13.399 "data_size": 65536 00:13:13.399 }, 00:13:13.399 { 00:13:13.399 "name": null, 00:13:13.399 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:13.399 "is_configured": false, 00:13:13.399 "data_offset": 0, 00:13:13.399 "data_size": 65536 00:13:13.399 }, 00:13:13.399 { 00:13:13.399 "name": "BaseBdev3", 00:13:13.399 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:13.399 "is_configured": true, 00:13:13.399 "data_offset": 0, 00:13:13.399 "data_size": 65536 00:13:13.399 } 00:13:13.399 ] 00:13:13.399 }' 00:13:13.399 16:51:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.399 16:51:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.659 [2024-09-29 16:51:35.277328] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:13.659 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.917 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.918 "name": "Existed_Raid", 00:13:13.918 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.918 "strip_size_kb": 64, 00:13:13.918 "state": "configuring", 00:13:13.918 "raid_level": "raid5f", 00:13:13.918 "superblock": false, 00:13:13.918 "num_base_bdevs": 3, 00:13:13.918 "num_base_bdevs_discovered": 1, 00:13:13.918 "num_base_bdevs_operational": 3, 00:13:13.918 "base_bdevs_list": [ 00:13:13.918 { 00:13:13.918 "name": null, 00:13:13.918 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:13.918 "is_configured": false, 00:13:13.918 "data_offset": 0, 00:13:13.918 "data_size": 65536 00:13:13.918 }, 00:13:13.918 { 00:13:13.918 "name": null, 00:13:13.918 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:13.918 "is_configured": false, 00:13:13.918 "data_offset": 0, 00:13:13.918 "data_size": 65536 00:13:13.918 }, 00:13:13.918 { 00:13:13.918 "name": "BaseBdev3", 00:13:13.918 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:13.918 "is_configured": true, 00:13:13.918 "data_offset": 0, 00:13:13.918 "data_size": 65536 00:13:13.918 } 00:13:13.918 ] 00:13:13.918 }' 00:13:13.918 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.918 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.176 [2024-09-29 16:51:35.779121] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.176 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.176 "name": "Existed_Raid", 00:13:14.177 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.177 "strip_size_kb": 64, 00:13:14.177 "state": "configuring", 00:13:14.177 "raid_level": "raid5f", 00:13:14.177 "superblock": false, 00:13:14.177 "num_base_bdevs": 3, 00:13:14.177 "num_base_bdevs_discovered": 2, 00:13:14.177 "num_base_bdevs_operational": 3, 00:13:14.177 "base_bdevs_list": [ 00:13:14.177 { 00:13:14.177 "name": null, 00:13:14.177 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:14.177 "is_configured": false, 00:13:14.177 "data_offset": 0, 00:13:14.177 "data_size": 65536 00:13:14.177 }, 00:13:14.177 { 00:13:14.177 "name": "BaseBdev2", 00:13:14.177 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:14.177 "is_configured": true, 00:13:14.177 "data_offset": 0, 00:13:14.177 "data_size": 65536 00:13:14.177 }, 00:13:14.177 { 00:13:14.177 "name": "BaseBdev3", 00:13:14.177 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:14.177 "is_configured": true, 00:13:14.177 "data_offset": 0, 00:13:14.177 "data_size": 65536 00:13:14.177 } 00:13:14.177 ] 00:13:14.177 }' 00:13:14.177 16:51:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.177 16:51:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 331f9cfb-d794-4513-b8eb-43940f90e306 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.745 [2024-09-29 16:51:36.293148] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:14.745 [2024-09-29 16:51:36.293244] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:14.745 [2024-09-29 16:51:36.293271] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:14.745 [2024-09-29 16:51:36.293543] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:14.745 [2024-09-29 16:51:36.293988] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:14.745 [2024-09-29 16:51:36.294040] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:14.745 [2024-09-29 16:51:36.294280] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:14.745 NewBaseBdev 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.745 [ 00:13:14.745 { 00:13:14.745 "name": "NewBaseBdev", 00:13:14.745 "aliases": [ 00:13:14.745 "331f9cfb-d794-4513-b8eb-43940f90e306" 00:13:14.745 ], 00:13:14.745 "product_name": "Malloc disk", 00:13:14.745 "block_size": 512, 00:13:14.745 "num_blocks": 65536, 00:13:14.745 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:14.745 "assigned_rate_limits": { 00:13:14.745 "rw_ios_per_sec": 0, 00:13:14.745 "rw_mbytes_per_sec": 0, 00:13:14.745 "r_mbytes_per_sec": 0, 00:13:14.745 "w_mbytes_per_sec": 0 00:13:14.745 }, 00:13:14.745 "claimed": true, 00:13:14.745 "claim_type": "exclusive_write", 00:13:14.745 "zoned": false, 00:13:14.745 "supported_io_types": { 00:13:14.745 "read": true, 00:13:14.745 "write": true, 00:13:14.745 "unmap": true, 00:13:14.745 "flush": true, 00:13:14.745 "reset": true, 00:13:14.745 "nvme_admin": false, 00:13:14.745 "nvme_io": false, 00:13:14.745 "nvme_io_md": false, 00:13:14.745 "write_zeroes": true, 00:13:14.745 "zcopy": true, 00:13:14.745 "get_zone_info": false, 00:13:14.745 "zone_management": false, 00:13:14.745 "zone_append": false, 00:13:14.745 "compare": false, 00:13:14.745 "compare_and_write": false, 00:13:14.745 "abort": true, 00:13:14.745 "seek_hole": false, 00:13:14.745 "seek_data": false, 00:13:14.745 "copy": true, 00:13:14.745 "nvme_iov_md": false 00:13:14.745 }, 00:13:14.745 "memory_domains": [ 00:13:14.745 { 00:13:14.745 "dma_device_id": "system", 00:13:14.745 "dma_device_type": 1 00:13:14.745 }, 00:13:14.745 { 00:13:14.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:14.745 "dma_device_type": 2 00:13:14.745 } 00:13:14.745 ], 00:13:14.745 "driver_specific": {} 00:13:14.745 } 00:13:14.745 ] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:14.745 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:14.746 "name": "Existed_Raid", 00:13:14.746 "uuid": "97f3f49c-8ad5-4383-833b-21000bbf79bd", 00:13:14.746 "strip_size_kb": 64, 00:13:14.746 "state": "online", 00:13:14.746 "raid_level": "raid5f", 00:13:14.746 "superblock": false, 00:13:14.746 "num_base_bdevs": 3, 00:13:14.746 "num_base_bdevs_discovered": 3, 00:13:14.746 "num_base_bdevs_operational": 3, 00:13:14.746 "base_bdevs_list": [ 00:13:14.746 { 00:13:14.746 "name": "NewBaseBdev", 00:13:14.746 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:14.746 "is_configured": true, 00:13:14.746 "data_offset": 0, 00:13:14.746 "data_size": 65536 00:13:14.746 }, 00:13:14.746 { 00:13:14.746 "name": "BaseBdev2", 00:13:14.746 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:14.746 "is_configured": true, 00:13:14.746 "data_offset": 0, 00:13:14.746 "data_size": 65536 00:13:14.746 }, 00:13:14.746 { 00:13:14.746 "name": "BaseBdev3", 00:13:14.746 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:14.746 "is_configured": true, 00:13:14.746 "data_offset": 0, 00:13:14.746 "data_size": 65536 00:13:14.746 } 00:13:14.746 ] 00:13:14.746 }' 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:14.746 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.314 [2024-09-29 16:51:36.788487] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:15.314 "name": "Existed_Raid", 00:13:15.314 "aliases": [ 00:13:15.314 "97f3f49c-8ad5-4383-833b-21000bbf79bd" 00:13:15.314 ], 00:13:15.314 "product_name": "Raid Volume", 00:13:15.314 "block_size": 512, 00:13:15.314 "num_blocks": 131072, 00:13:15.314 "uuid": "97f3f49c-8ad5-4383-833b-21000bbf79bd", 00:13:15.314 "assigned_rate_limits": { 00:13:15.314 "rw_ios_per_sec": 0, 00:13:15.314 "rw_mbytes_per_sec": 0, 00:13:15.314 "r_mbytes_per_sec": 0, 00:13:15.314 "w_mbytes_per_sec": 0 00:13:15.314 }, 00:13:15.314 "claimed": false, 00:13:15.314 "zoned": false, 00:13:15.314 "supported_io_types": { 00:13:15.314 "read": true, 00:13:15.314 "write": true, 00:13:15.314 "unmap": false, 00:13:15.314 "flush": false, 00:13:15.314 "reset": true, 00:13:15.314 "nvme_admin": false, 00:13:15.314 "nvme_io": false, 00:13:15.314 "nvme_io_md": false, 00:13:15.314 "write_zeroes": true, 00:13:15.314 "zcopy": false, 00:13:15.314 "get_zone_info": false, 00:13:15.314 "zone_management": false, 00:13:15.314 "zone_append": false, 00:13:15.314 "compare": false, 00:13:15.314 "compare_and_write": false, 00:13:15.314 "abort": false, 00:13:15.314 "seek_hole": false, 00:13:15.314 "seek_data": false, 00:13:15.314 "copy": false, 00:13:15.314 "nvme_iov_md": false 00:13:15.314 }, 00:13:15.314 "driver_specific": { 00:13:15.314 "raid": { 00:13:15.314 "uuid": "97f3f49c-8ad5-4383-833b-21000bbf79bd", 00:13:15.314 "strip_size_kb": 64, 00:13:15.314 "state": "online", 00:13:15.314 "raid_level": "raid5f", 00:13:15.314 "superblock": false, 00:13:15.314 "num_base_bdevs": 3, 00:13:15.314 "num_base_bdevs_discovered": 3, 00:13:15.314 "num_base_bdevs_operational": 3, 00:13:15.314 "base_bdevs_list": [ 00:13:15.314 { 00:13:15.314 "name": "NewBaseBdev", 00:13:15.314 "uuid": "331f9cfb-d794-4513-b8eb-43940f90e306", 00:13:15.314 "is_configured": true, 00:13:15.314 "data_offset": 0, 00:13:15.314 "data_size": 65536 00:13:15.314 }, 00:13:15.314 { 00:13:15.314 "name": "BaseBdev2", 00:13:15.314 "uuid": "4f054a1d-fba6-4552-bfa9-9319854bd7a5", 00:13:15.314 "is_configured": true, 00:13:15.314 "data_offset": 0, 00:13:15.314 "data_size": 65536 00:13:15.314 }, 00:13:15.314 { 00:13:15.314 "name": "BaseBdev3", 00:13:15.314 "uuid": "26dfe549-8423-42f6-b123-bb7d750e2524", 00:13:15.314 "is_configured": true, 00:13:15.314 "data_offset": 0, 00:13:15.314 "data_size": 65536 00:13:15.314 } 00:13:15.314 ] 00:13:15.314 } 00:13:15.314 } 00:13:15.314 }' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:15.314 BaseBdev2 00:13:15.314 BaseBdev3' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.314 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.574 16:51:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.574 [2024-09-29 16:51:37.059828] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:15.574 [2024-09-29 16:51:37.059894] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:15.574 [2024-09-29 16:51:37.059976] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:15.574 [2024-09-29 16:51:37.060240] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:15.574 [2024-09-29 16:51:37.060296] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90094 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 90094 ']' 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 90094 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90094 00:13:15.574 killing process with pid 90094 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90094' 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 90094 00:13:15.574 [2024-09-29 16:51:37.106120] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:15.574 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 90094 00:13:15.574 [2024-09-29 16:51:37.137111] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:15.834 00:13:15.834 real 0m9.125s 00:13:15.834 user 0m15.496s 00:13:15.834 sys 0m2.007s 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:15.834 ************************************ 00:13:15.834 END TEST raid5f_state_function_test 00:13:15.834 ************************************ 00:13:15.834 16:51:37 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:13:15.834 16:51:37 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:13:15.834 16:51:37 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:15.834 16:51:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:15.834 ************************************ 00:13:15.834 START TEST raid5f_state_function_test_sb 00:13:15.834 ************************************ 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 true 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:15.834 Process raid pid: 90699 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=90699 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90699' 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 90699 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 90699 ']' 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:15.834 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:15.834 16:51:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.094 [2024-09-29 16:51:37.571677] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:13:16.094 [2024-09-29 16:51:37.572551] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:16.094 [2024-09-29 16:51:37.721143] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:16.353 [2024-09-29 16:51:37.767940] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:16.353 [2024-09-29 16:51:37.810975] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:16.353 [2024-09-29 16:51:37.811106] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.922 [2024-09-29 16:51:38.388760] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:16.922 [2024-09-29 16:51:38.388883] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:16.922 [2024-09-29 16:51:38.388915] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:16.922 [2024-09-29 16:51:38.388938] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:16.922 [2024-09-29 16:51:38.388955] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:16.922 [2024-09-29 16:51:38.388979] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.922 "name": "Existed_Raid", 00:13:16.922 "uuid": "8c6c1936-0dcf-4ae9-b1ba-0b1990d1946c", 00:13:16.922 "strip_size_kb": 64, 00:13:16.922 "state": "configuring", 00:13:16.922 "raid_level": "raid5f", 00:13:16.922 "superblock": true, 00:13:16.922 "num_base_bdevs": 3, 00:13:16.922 "num_base_bdevs_discovered": 0, 00:13:16.922 "num_base_bdevs_operational": 3, 00:13:16.922 "base_bdevs_list": [ 00:13:16.922 { 00:13:16.922 "name": "BaseBdev1", 00:13:16.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.922 "is_configured": false, 00:13:16.922 "data_offset": 0, 00:13:16.922 "data_size": 0 00:13:16.922 }, 00:13:16.922 { 00:13:16.922 "name": "BaseBdev2", 00:13:16.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.922 "is_configured": false, 00:13:16.922 "data_offset": 0, 00:13:16.922 "data_size": 0 00:13:16.922 }, 00:13:16.922 { 00:13:16.922 "name": "BaseBdev3", 00:13:16.922 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.922 "is_configured": false, 00:13:16.922 "data_offset": 0, 00:13:16.922 "data_size": 0 00:13:16.922 } 00:13:16.922 ] 00:13:16.922 }' 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.922 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.182 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:17.182 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.182 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.182 [2024-09-29 16:51:38.791924] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:17.182 [2024-09-29 16:51:38.792031] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:17.182 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.183 [2024-09-29 16:51:38.803944] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:17.183 [2024-09-29 16:51:38.804027] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:17.183 [2024-09-29 16:51:38.804065] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:17.183 [2024-09-29 16:51:38.804106] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:17.183 [2024-09-29 16:51:38.804124] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:17.183 [2024-09-29 16:51:38.804145] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.183 [2024-09-29 16:51:38.824806] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:17.183 BaseBdev1 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.183 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.183 [ 00:13:17.183 { 00:13:17.183 "name": "BaseBdev1", 00:13:17.183 "aliases": [ 00:13:17.183 "29aff16d-678b-4860-85e3-c7acbf52bb14" 00:13:17.183 ], 00:13:17.183 "product_name": "Malloc disk", 00:13:17.183 "block_size": 512, 00:13:17.183 "num_blocks": 65536, 00:13:17.183 "uuid": "29aff16d-678b-4860-85e3-c7acbf52bb14", 00:13:17.183 "assigned_rate_limits": { 00:13:17.183 "rw_ios_per_sec": 0, 00:13:17.183 "rw_mbytes_per_sec": 0, 00:13:17.183 "r_mbytes_per_sec": 0, 00:13:17.183 "w_mbytes_per_sec": 0 00:13:17.183 }, 00:13:17.183 "claimed": true, 00:13:17.183 "claim_type": "exclusive_write", 00:13:17.183 "zoned": false, 00:13:17.183 "supported_io_types": { 00:13:17.183 "read": true, 00:13:17.183 "write": true, 00:13:17.183 "unmap": true, 00:13:17.183 "flush": true, 00:13:17.442 "reset": true, 00:13:17.442 "nvme_admin": false, 00:13:17.442 "nvme_io": false, 00:13:17.442 "nvme_io_md": false, 00:13:17.442 "write_zeroes": true, 00:13:17.442 "zcopy": true, 00:13:17.442 "get_zone_info": false, 00:13:17.442 "zone_management": false, 00:13:17.442 "zone_append": false, 00:13:17.442 "compare": false, 00:13:17.442 "compare_and_write": false, 00:13:17.442 "abort": true, 00:13:17.442 "seek_hole": false, 00:13:17.442 "seek_data": false, 00:13:17.442 "copy": true, 00:13:17.442 "nvme_iov_md": false 00:13:17.442 }, 00:13:17.442 "memory_domains": [ 00:13:17.442 { 00:13:17.442 "dma_device_id": "system", 00:13:17.442 "dma_device_type": 1 00:13:17.442 }, 00:13:17.442 { 00:13:17.442 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:17.442 "dma_device_type": 2 00:13:17.442 } 00:13:17.442 ], 00:13:17.442 "driver_specific": {} 00:13:17.442 } 00:13:17.442 ] 00:13:17.442 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.442 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:17.443 "name": "Existed_Raid", 00:13:17.443 "uuid": "cc9cfdb9-2a8c-421d-b7f5-0c28ad3e22d5", 00:13:17.443 "strip_size_kb": 64, 00:13:17.443 "state": "configuring", 00:13:17.443 "raid_level": "raid5f", 00:13:17.443 "superblock": true, 00:13:17.443 "num_base_bdevs": 3, 00:13:17.443 "num_base_bdevs_discovered": 1, 00:13:17.443 "num_base_bdevs_operational": 3, 00:13:17.443 "base_bdevs_list": [ 00:13:17.443 { 00:13:17.443 "name": "BaseBdev1", 00:13:17.443 "uuid": "29aff16d-678b-4860-85e3-c7acbf52bb14", 00:13:17.443 "is_configured": true, 00:13:17.443 "data_offset": 2048, 00:13:17.443 "data_size": 63488 00:13:17.443 }, 00:13:17.443 { 00:13:17.443 "name": "BaseBdev2", 00:13:17.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.443 "is_configured": false, 00:13:17.443 "data_offset": 0, 00:13:17.443 "data_size": 0 00:13:17.443 }, 00:13:17.443 { 00:13:17.443 "name": "BaseBdev3", 00:13:17.443 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.443 "is_configured": false, 00:13:17.443 "data_offset": 0, 00:13:17.443 "data_size": 0 00:13:17.443 } 00:13:17.443 ] 00:13:17.443 }' 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:17.443 16:51:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.702 [2024-09-29 16:51:39.327962] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:17.702 [2024-09-29 16:51:39.328077] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.702 [2024-09-29 16:51:39.340000] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:17.702 [2024-09-29 16:51:39.341874] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:17.702 [2024-09-29 16:51:39.341964] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:17.702 [2024-09-29 16:51:39.341990] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:17.702 [2024-09-29 16:51:39.342013] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.702 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.962 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:17.962 "name": "Existed_Raid", 00:13:17.962 "uuid": "03b65507-710c-44ec-982e-00e935705fd6", 00:13:17.962 "strip_size_kb": 64, 00:13:17.962 "state": "configuring", 00:13:17.962 "raid_level": "raid5f", 00:13:17.962 "superblock": true, 00:13:17.962 "num_base_bdevs": 3, 00:13:17.962 "num_base_bdevs_discovered": 1, 00:13:17.962 "num_base_bdevs_operational": 3, 00:13:17.962 "base_bdevs_list": [ 00:13:17.962 { 00:13:17.962 "name": "BaseBdev1", 00:13:17.962 "uuid": "29aff16d-678b-4860-85e3-c7acbf52bb14", 00:13:17.962 "is_configured": true, 00:13:17.962 "data_offset": 2048, 00:13:17.962 "data_size": 63488 00:13:17.962 }, 00:13:17.962 { 00:13:17.962 "name": "BaseBdev2", 00:13:17.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.962 "is_configured": false, 00:13:17.962 "data_offset": 0, 00:13:17.962 "data_size": 0 00:13:17.962 }, 00:13:17.962 { 00:13:17.962 "name": "BaseBdev3", 00:13:17.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.962 "is_configured": false, 00:13:17.962 "data_offset": 0, 00:13:17.962 "data_size": 0 00:13:17.962 } 00:13:17.962 ] 00:13:17.962 }' 00:13:17.962 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:17.962 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.222 [2024-09-29 16:51:39.836563] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:18.222 BaseBdev2 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.222 [ 00:13:18.222 { 00:13:18.222 "name": "BaseBdev2", 00:13:18.222 "aliases": [ 00:13:18.222 "62e1f8f0-c56a-43ee-8f34-bdcd51eac925" 00:13:18.222 ], 00:13:18.222 "product_name": "Malloc disk", 00:13:18.222 "block_size": 512, 00:13:18.222 "num_blocks": 65536, 00:13:18.222 "uuid": "62e1f8f0-c56a-43ee-8f34-bdcd51eac925", 00:13:18.222 "assigned_rate_limits": { 00:13:18.222 "rw_ios_per_sec": 0, 00:13:18.222 "rw_mbytes_per_sec": 0, 00:13:18.222 "r_mbytes_per_sec": 0, 00:13:18.222 "w_mbytes_per_sec": 0 00:13:18.222 }, 00:13:18.222 "claimed": true, 00:13:18.222 "claim_type": "exclusive_write", 00:13:18.222 "zoned": false, 00:13:18.222 "supported_io_types": { 00:13:18.222 "read": true, 00:13:18.222 "write": true, 00:13:18.222 "unmap": true, 00:13:18.222 "flush": true, 00:13:18.222 "reset": true, 00:13:18.222 "nvme_admin": false, 00:13:18.222 "nvme_io": false, 00:13:18.222 "nvme_io_md": false, 00:13:18.222 "write_zeroes": true, 00:13:18.222 "zcopy": true, 00:13:18.222 "get_zone_info": false, 00:13:18.222 "zone_management": false, 00:13:18.222 "zone_append": false, 00:13:18.222 "compare": false, 00:13:18.222 "compare_and_write": false, 00:13:18.222 "abort": true, 00:13:18.222 "seek_hole": false, 00:13:18.222 "seek_data": false, 00:13:18.222 "copy": true, 00:13:18.222 "nvme_iov_md": false 00:13:18.222 }, 00:13:18.222 "memory_domains": [ 00:13:18.222 { 00:13:18.222 "dma_device_id": "system", 00:13:18.222 "dma_device_type": 1 00:13:18.222 }, 00:13:18.222 { 00:13:18.222 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:18.222 "dma_device_type": 2 00:13:18.222 } 00:13:18.222 ], 00:13:18.222 "driver_specific": {} 00:13:18.222 } 00:13:18.222 ] 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.222 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.482 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.482 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:18.482 "name": "Existed_Raid", 00:13:18.482 "uuid": "03b65507-710c-44ec-982e-00e935705fd6", 00:13:18.482 "strip_size_kb": 64, 00:13:18.482 "state": "configuring", 00:13:18.482 "raid_level": "raid5f", 00:13:18.482 "superblock": true, 00:13:18.482 "num_base_bdevs": 3, 00:13:18.482 "num_base_bdevs_discovered": 2, 00:13:18.482 "num_base_bdevs_operational": 3, 00:13:18.482 "base_bdevs_list": [ 00:13:18.482 { 00:13:18.482 "name": "BaseBdev1", 00:13:18.482 "uuid": "29aff16d-678b-4860-85e3-c7acbf52bb14", 00:13:18.482 "is_configured": true, 00:13:18.482 "data_offset": 2048, 00:13:18.482 "data_size": 63488 00:13:18.482 }, 00:13:18.482 { 00:13:18.482 "name": "BaseBdev2", 00:13:18.482 "uuid": "62e1f8f0-c56a-43ee-8f34-bdcd51eac925", 00:13:18.482 "is_configured": true, 00:13:18.482 "data_offset": 2048, 00:13:18.482 "data_size": 63488 00:13:18.482 }, 00:13:18.482 { 00:13:18.482 "name": "BaseBdev3", 00:13:18.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:18.482 "is_configured": false, 00:13:18.482 "data_offset": 0, 00:13:18.482 "data_size": 0 00:13:18.482 } 00:13:18.482 ] 00:13:18.482 }' 00:13:18.482 16:51:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:18.482 16:51:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.742 [2024-09-29 16:51:40.370510] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:18.742 [2024-09-29 16:51:40.370825] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:18.742 [2024-09-29 16:51:40.370885] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:18.742 BaseBdev3 00:13:18.742 [2024-09-29 16:51:40.371234] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:18.742 [2024-09-29 16:51:40.371681] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:18.742 [2024-09-29 16:51:40.371757] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:18.742 [2024-09-29 16:51:40.371913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:18.742 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:18.743 [ 00:13:18.743 { 00:13:18.743 "name": "BaseBdev3", 00:13:18.743 "aliases": [ 00:13:18.743 "aaea16c4-e441-46fa-9d22-00996e9be2b3" 00:13:18.743 ], 00:13:18.743 "product_name": "Malloc disk", 00:13:18.743 "block_size": 512, 00:13:18.743 "num_blocks": 65536, 00:13:18.743 "uuid": "aaea16c4-e441-46fa-9d22-00996e9be2b3", 00:13:18.743 "assigned_rate_limits": { 00:13:18.743 "rw_ios_per_sec": 0, 00:13:18.743 "rw_mbytes_per_sec": 0, 00:13:18.743 "r_mbytes_per_sec": 0, 00:13:18.743 "w_mbytes_per_sec": 0 00:13:18.743 }, 00:13:18.743 "claimed": true, 00:13:18.743 "claim_type": "exclusive_write", 00:13:18.743 "zoned": false, 00:13:18.743 "supported_io_types": { 00:13:18.743 "read": true, 00:13:18.743 "write": true, 00:13:18.743 "unmap": true, 00:13:18.743 "flush": true, 00:13:18.743 "reset": true, 00:13:18.743 "nvme_admin": false, 00:13:18.743 "nvme_io": false, 00:13:18.743 "nvme_io_md": false, 00:13:18.743 "write_zeroes": true, 00:13:18.743 "zcopy": true, 00:13:18.743 "get_zone_info": false, 00:13:18.743 "zone_management": false, 00:13:18.743 "zone_append": false, 00:13:18.743 "compare": false, 00:13:18.743 "compare_and_write": false, 00:13:18.743 "abort": true, 00:13:18.743 "seek_hole": false, 00:13:18.743 "seek_data": false, 00:13:18.743 "copy": true, 00:13:18.743 "nvme_iov_md": false 00:13:18.743 }, 00:13:18.743 "memory_domains": [ 00:13:18.743 { 00:13:18.743 "dma_device_id": "system", 00:13:18.743 "dma_device_type": 1 00:13:18.743 }, 00:13:18.743 { 00:13:18.743 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:18.743 "dma_device_type": 2 00:13:18.743 } 00:13:18.743 ], 00:13:18.743 "driver_specific": {} 00:13:18.743 } 00:13:18.743 ] 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:18.743 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.003 "name": "Existed_Raid", 00:13:19.003 "uuid": "03b65507-710c-44ec-982e-00e935705fd6", 00:13:19.003 "strip_size_kb": 64, 00:13:19.003 "state": "online", 00:13:19.003 "raid_level": "raid5f", 00:13:19.003 "superblock": true, 00:13:19.003 "num_base_bdevs": 3, 00:13:19.003 "num_base_bdevs_discovered": 3, 00:13:19.003 "num_base_bdevs_operational": 3, 00:13:19.003 "base_bdevs_list": [ 00:13:19.003 { 00:13:19.003 "name": "BaseBdev1", 00:13:19.003 "uuid": "29aff16d-678b-4860-85e3-c7acbf52bb14", 00:13:19.003 "is_configured": true, 00:13:19.003 "data_offset": 2048, 00:13:19.003 "data_size": 63488 00:13:19.003 }, 00:13:19.003 { 00:13:19.003 "name": "BaseBdev2", 00:13:19.003 "uuid": "62e1f8f0-c56a-43ee-8f34-bdcd51eac925", 00:13:19.003 "is_configured": true, 00:13:19.003 "data_offset": 2048, 00:13:19.003 "data_size": 63488 00:13:19.003 }, 00:13:19.003 { 00:13:19.003 "name": "BaseBdev3", 00:13:19.003 "uuid": "aaea16c4-e441-46fa-9d22-00996e9be2b3", 00:13:19.003 "is_configured": true, 00:13:19.003 "data_offset": 2048, 00:13:19.003 "data_size": 63488 00:13:19.003 } 00:13:19.003 ] 00:13:19.003 }' 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.003 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.262 [2024-09-29 16:51:40.893853] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.262 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:19.262 "name": "Existed_Raid", 00:13:19.262 "aliases": [ 00:13:19.262 "03b65507-710c-44ec-982e-00e935705fd6" 00:13:19.262 ], 00:13:19.262 "product_name": "Raid Volume", 00:13:19.263 "block_size": 512, 00:13:19.263 "num_blocks": 126976, 00:13:19.263 "uuid": "03b65507-710c-44ec-982e-00e935705fd6", 00:13:19.263 "assigned_rate_limits": { 00:13:19.263 "rw_ios_per_sec": 0, 00:13:19.263 "rw_mbytes_per_sec": 0, 00:13:19.263 "r_mbytes_per_sec": 0, 00:13:19.263 "w_mbytes_per_sec": 0 00:13:19.263 }, 00:13:19.263 "claimed": false, 00:13:19.263 "zoned": false, 00:13:19.263 "supported_io_types": { 00:13:19.263 "read": true, 00:13:19.263 "write": true, 00:13:19.263 "unmap": false, 00:13:19.263 "flush": false, 00:13:19.263 "reset": true, 00:13:19.263 "nvme_admin": false, 00:13:19.263 "nvme_io": false, 00:13:19.263 "nvme_io_md": false, 00:13:19.263 "write_zeroes": true, 00:13:19.263 "zcopy": false, 00:13:19.263 "get_zone_info": false, 00:13:19.263 "zone_management": false, 00:13:19.263 "zone_append": false, 00:13:19.263 "compare": false, 00:13:19.263 "compare_and_write": false, 00:13:19.263 "abort": false, 00:13:19.263 "seek_hole": false, 00:13:19.263 "seek_data": false, 00:13:19.263 "copy": false, 00:13:19.263 "nvme_iov_md": false 00:13:19.263 }, 00:13:19.263 "driver_specific": { 00:13:19.263 "raid": { 00:13:19.263 "uuid": "03b65507-710c-44ec-982e-00e935705fd6", 00:13:19.263 "strip_size_kb": 64, 00:13:19.263 "state": "online", 00:13:19.263 "raid_level": "raid5f", 00:13:19.263 "superblock": true, 00:13:19.263 "num_base_bdevs": 3, 00:13:19.263 "num_base_bdevs_discovered": 3, 00:13:19.263 "num_base_bdevs_operational": 3, 00:13:19.263 "base_bdevs_list": [ 00:13:19.263 { 00:13:19.263 "name": "BaseBdev1", 00:13:19.263 "uuid": "29aff16d-678b-4860-85e3-c7acbf52bb14", 00:13:19.263 "is_configured": true, 00:13:19.263 "data_offset": 2048, 00:13:19.263 "data_size": 63488 00:13:19.263 }, 00:13:19.263 { 00:13:19.263 "name": "BaseBdev2", 00:13:19.263 "uuid": "62e1f8f0-c56a-43ee-8f34-bdcd51eac925", 00:13:19.263 "is_configured": true, 00:13:19.263 "data_offset": 2048, 00:13:19.263 "data_size": 63488 00:13:19.263 }, 00:13:19.263 { 00:13:19.263 "name": "BaseBdev3", 00:13:19.263 "uuid": "aaea16c4-e441-46fa-9d22-00996e9be2b3", 00:13:19.263 "is_configured": true, 00:13:19.263 "data_offset": 2048, 00:13:19.263 "data_size": 63488 00:13:19.263 } 00:13:19.263 ] 00:13:19.263 } 00:13:19.263 } 00:13:19.263 }' 00:13:19.263 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:19.523 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:19.523 BaseBdev2 00:13:19.523 BaseBdev3' 00:13:19.523 16:51:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.523 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.783 [2024-09-29 16:51:41.197188] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.783 "name": "Existed_Raid", 00:13:19.783 "uuid": "03b65507-710c-44ec-982e-00e935705fd6", 00:13:19.783 "strip_size_kb": 64, 00:13:19.783 "state": "online", 00:13:19.783 "raid_level": "raid5f", 00:13:19.783 "superblock": true, 00:13:19.783 "num_base_bdevs": 3, 00:13:19.783 "num_base_bdevs_discovered": 2, 00:13:19.783 "num_base_bdevs_operational": 2, 00:13:19.783 "base_bdevs_list": [ 00:13:19.783 { 00:13:19.783 "name": null, 00:13:19.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.783 "is_configured": false, 00:13:19.783 "data_offset": 0, 00:13:19.783 "data_size": 63488 00:13:19.783 }, 00:13:19.783 { 00:13:19.783 "name": "BaseBdev2", 00:13:19.783 "uuid": "62e1f8f0-c56a-43ee-8f34-bdcd51eac925", 00:13:19.783 "is_configured": true, 00:13:19.783 "data_offset": 2048, 00:13:19.783 "data_size": 63488 00:13:19.783 }, 00:13:19.783 { 00:13:19.783 "name": "BaseBdev3", 00:13:19.783 "uuid": "aaea16c4-e441-46fa-9d22-00996e9be2b3", 00:13:19.783 "is_configured": true, 00:13:19.783 "data_offset": 2048, 00:13:19.783 "data_size": 63488 00:13:19.783 } 00:13:19.783 ] 00:13:19.783 }' 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.783 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.042 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:20.042 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:20.042 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.043 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.303 [2024-09-29 16:51:41.715869] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:20.303 [2024-09-29 16:51:41.716043] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:20.303 [2024-09-29 16:51:41.727390] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.303 [2024-09-29 16:51:41.783334] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:20.303 [2024-09-29 16:51:41.783423] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.303 BaseBdev2 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.303 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.303 [ 00:13:20.304 { 00:13:20.304 "name": "BaseBdev2", 00:13:20.304 "aliases": [ 00:13:20.304 "f4c0ac8d-4181-41f7-94e2-993c724ba2d4" 00:13:20.304 ], 00:13:20.304 "product_name": "Malloc disk", 00:13:20.304 "block_size": 512, 00:13:20.304 "num_blocks": 65536, 00:13:20.304 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:20.304 "assigned_rate_limits": { 00:13:20.304 "rw_ios_per_sec": 0, 00:13:20.304 "rw_mbytes_per_sec": 0, 00:13:20.304 "r_mbytes_per_sec": 0, 00:13:20.304 "w_mbytes_per_sec": 0 00:13:20.304 }, 00:13:20.304 "claimed": false, 00:13:20.304 "zoned": false, 00:13:20.304 "supported_io_types": { 00:13:20.304 "read": true, 00:13:20.304 "write": true, 00:13:20.304 "unmap": true, 00:13:20.304 "flush": true, 00:13:20.304 "reset": true, 00:13:20.304 "nvme_admin": false, 00:13:20.304 "nvme_io": false, 00:13:20.304 "nvme_io_md": false, 00:13:20.304 "write_zeroes": true, 00:13:20.304 "zcopy": true, 00:13:20.304 "get_zone_info": false, 00:13:20.304 "zone_management": false, 00:13:20.304 "zone_append": false, 00:13:20.304 "compare": false, 00:13:20.304 "compare_and_write": false, 00:13:20.304 "abort": true, 00:13:20.304 "seek_hole": false, 00:13:20.304 "seek_data": false, 00:13:20.304 "copy": true, 00:13:20.304 "nvme_iov_md": false 00:13:20.304 }, 00:13:20.304 "memory_domains": [ 00:13:20.304 { 00:13:20.304 "dma_device_id": "system", 00:13:20.304 "dma_device_type": 1 00:13:20.304 }, 00:13:20.304 { 00:13:20.304 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:20.304 "dma_device_type": 2 00:13:20.304 } 00:13:20.304 ], 00:13:20.304 "driver_specific": {} 00:13:20.304 } 00:13:20.304 ] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.304 BaseBdev3 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.304 [ 00:13:20.304 { 00:13:20.304 "name": "BaseBdev3", 00:13:20.304 "aliases": [ 00:13:20.304 "37a6ef20-9186-4047-b710-af81de7486be" 00:13:20.304 ], 00:13:20.304 "product_name": "Malloc disk", 00:13:20.304 "block_size": 512, 00:13:20.304 "num_blocks": 65536, 00:13:20.304 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:20.304 "assigned_rate_limits": { 00:13:20.304 "rw_ios_per_sec": 0, 00:13:20.304 "rw_mbytes_per_sec": 0, 00:13:20.304 "r_mbytes_per_sec": 0, 00:13:20.304 "w_mbytes_per_sec": 0 00:13:20.304 }, 00:13:20.304 "claimed": false, 00:13:20.304 "zoned": false, 00:13:20.304 "supported_io_types": { 00:13:20.304 "read": true, 00:13:20.304 "write": true, 00:13:20.304 "unmap": true, 00:13:20.304 "flush": true, 00:13:20.304 "reset": true, 00:13:20.304 "nvme_admin": false, 00:13:20.304 "nvme_io": false, 00:13:20.304 "nvme_io_md": false, 00:13:20.304 "write_zeroes": true, 00:13:20.304 "zcopy": true, 00:13:20.304 "get_zone_info": false, 00:13:20.304 "zone_management": false, 00:13:20.304 "zone_append": false, 00:13:20.304 "compare": false, 00:13:20.304 "compare_and_write": false, 00:13:20.304 "abort": true, 00:13:20.304 "seek_hole": false, 00:13:20.304 "seek_data": false, 00:13:20.304 "copy": true, 00:13:20.304 "nvme_iov_md": false 00:13:20.304 }, 00:13:20.304 "memory_domains": [ 00:13:20.304 { 00:13:20.304 "dma_device_id": "system", 00:13:20.304 "dma_device_type": 1 00:13:20.304 }, 00:13:20.304 { 00:13:20.304 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:20.304 "dma_device_type": 2 00:13:20.304 } 00:13:20.304 ], 00:13:20.304 "driver_specific": {} 00:13:20.304 } 00:13:20.304 ] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.304 [2024-09-29 16:51:41.961983] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:20.304 [2024-09-29 16:51:41.962094] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:20.304 [2024-09-29 16:51:41.962132] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:20.304 [2024-09-29 16:51:41.964015] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:20.304 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.564 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.564 16:51:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.564 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:20.564 "name": "Existed_Raid", 00:13:20.564 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:20.564 "strip_size_kb": 64, 00:13:20.564 "state": "configuring", 00:13:20.564 "raid_level": "raid5f", 00:13:20.564 "superblock": true, 00:13:20.564 "num_base_bdevs": 3, 00:13:20.564 "num_base_bdevs_discovered": 2, 00:13:20.564 "num_base_bdevs_operational": 3, 00:13:20.564 "base_bdevs_list": [ 00:13:20.564 { 00:13:20.564 "name": "BaseBdev1", 00:13:20.564 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.564 "is_configured": false, 00:13:20.564 "data_offset": 0, 00:13:20.564 "data_size": 0 00:13:20.564 }, 00:13:20.564 { 00:13:20.564 "name": "BaseBdev2", 00:13:20.564 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:20.564 "is_configured": true, 00:13:20.564 "data_offset": 2048, 00:13:20.564 "data_size": 63488 00:13:20.564 }, 00:13:20.564 { 00:13:20.564 "name": "BaseBdev3", 00:13:20.564 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:20.564 "is_configured": true, 00:13:20.564 "data_offset": 2048, 00:13:20.564 "data_size": 63488 00:13:20.564 } 00:13:20.564 ] 00:13:20.564 }' 00:13:20.564 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:20.564 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.824 [2024-09-29 16:51:42.417217] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:20.824 "name": "Existed_Raid", 00:13:20.824 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:20.824 "strip_size_kb": 64, 00:13:20.824 "state": "configuring", 00:13:20.824 "raid_level": "raid5f", 00:13:20.824 "superblock": true, 00:13:20.824 "num_base_bdevs": 3, 00:13:20.824 "num_base_bdevs_discovered": 1, 00:13:20.824 "num_base_bdevs_operational": 3, 00:13:20.824 "base_bdevs_list": [ 00:13:20.824 { 00:13:20.824 "name": "BaseBdev1", 00:13:20.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:20.824 "is_configured": false, 00:13:20.824 "data_offset": 0, 00:13:20.824 "data_size": 0 00:13:20.824 }, 00:13:20.824 { 00:13:20.824 "name": null, 00:13:20.824 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:20.824 "is_configured": false, 00:13:20.824 "data_offset": 0, 00:13:20.824 "data_size": 63488 00:13:20.824 }, 00:13:20.824 { 00:13:20.824 "name": "BaseBdev3", 00:13:20.824 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:20.824 "is_configured": true, 00:13:20.824 "data_offset": 2048, 00:13:20.824 "data_size": 63488 00:13:20.824 } 00:13:20.824 ] 00:13:20.824 }' 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:20.824 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.394 [2024-09-29 16:51:42.975263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:21.394 BaseBdev1 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.394 16:51:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.394 [ 00:13:21.394 { 00:13:21.394 "name": "BaseBdev1", 00:13:21.394 "aliases": [ 00:13:21.394 "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4" 00:13:21.394 ], 00:13:21.394 "product_name": "Malloc disk", 00:13:21.394 "block_size": 512, 00:13:21.394 "num_blocks": 65536, 00:13:21.394 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:21.394 "assigned_rate_limits": { 00:13:21.394 "rw_ios_per_sec": 0, 00:13:21.394 "rw_mbytes_per_sec": 0, 00:13:21.394 "r_mbytes_per_sec": 0, 00:13:21.394 "w_mbytes_per_sec": 0 00:13:21.394 }, 00:13:21.394 "claimed": true, 00:13:21.394 "claim_type": "exclusive_write", 00:13:21.394 "zoned": false, 00:13:21.394 "supported_io_types": { 00:13:21.394 "read": true, 00:13:21.394 "write": true, 00:13:21.394 "unmap": true, 00:13:21.394 "flush": true, 00:13:21.394 "reset": true, 00:13:21.394 "nvme_admin": false, 00:13:21.394 "nvme_io": false, 00:13:21.394 "nvme_io_md": false, 00:13:21.394 "write_zeroes": true, 00:13:21.394 "zcopy": true, 00:13:21.394 "get_zone_info": false, 00:13:21.394 "zone_management": false, 00:13:21.394 "zone_append": false, 00:13:21.394 "compare": false, 00:13:21.394 "compare_and_write": false, 00:13:21.394 "abort": true, 00:13:21.394 "seek_hole": false, 00:13:21.394 "seek_data": false, 00:13:21.394 "copy": true, 00:13:21.394 "nvme_iov_md": false 00:13:21.394 }, 00:13:21.394 "memory_domains": [ 00:13:21.394 { 00:13:21.394 "dma_device_id": "system", 00:13:21.394 "dma_device_type": 1 00:13:21.394 }, 00:13:21.394 { 00:13:21.394 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:21.394 "dma_device_type": 2 00:13:21.394 } 00:13:21.394 ], 00:13:21.394 "driver_specific": {} 00:13:21.394 } 00:13:21.394 ] 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.394 "name": "Existed_Raid", 00:13:21.394 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:21.394 "strip_size_kb": 64, 00:13:21.394 "state": "configuring", 00:13:21.394 "raid_level": "raid5f", 00:13:21.394 "superblock": true, 00:13:21.394 "num_base_bdevs": 3, 00:13:21.394 "num_base_bdevs_discovered": 2, 00:13:21.394 "num_base_bdevs_operational": 3, 00:13:21.394 "base_bdevs_list": [ 00:13:21.394 { 00:13:21.394 "name": "BaseBdev1", 00:13:21.394 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:21.394 "is_configured": true, 00:13:21.394 "data_offset": 2048, 00:13:21.394 "data_size": 63488 00:13:21.394 }, 00:13:21.394 { 00:13:21.394 "name": null, 00:13:21.394 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:21.394 "is_configured": false, 00:13:21.394 "data_offset": 0, 00:13:21.394 "data_size": 63488 00:13:21.394 }, 00:13:21.394 { 00:13:21.394 "name": "BaseBdev3", 00:13:21.394 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:21.394 "is_configured": true, 00:13:21.394 "data_offset": 2048, 00:13:21.394 "data_size": 63488 00:13:21.394 } 00:13:21.394 ] 00:13:21.394 }' 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.394 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.964 [2024-09-29 16:51:43.494459] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.964 "name": "Existed_Raid", 00:13:21.964 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:21.964 "strip_size_kb": 64, 00:13:21.964 "state": "configuring", 00:13:21.964 "raid_level": "raid5f", 00:13:21.964 "superblock": true, 00:13:21.964 "num_base_bdevs": 3, 00:13:21.964 "num_base_bdevs_discovered": 1, 00:13:21.964 "num_base_bdevs_operational": 3, 00:13:21.964 "base_bdevs_list": [ 00:13:21.964 { 00:13:21.964 "name": "BaseBdev1", 00:13:21.964 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:21.964 "is_configured": true, 00:13:21.964 "data_offset": 2048, 00:13:21.964 "data_size": 63488 00:13:21.964 }, 00:13:21.964 { 00:13:21.964 "name": null, 00:13:21.964 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:21.964 "is_configured": false, 00:13:21.964 "data_offset": 0, 00:13:21.964 "data_size": 63488 00:13:21.964 }, 00:13:21.964 { 00:13:21.964 "name": null, 00:13:21.964 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:21.964 "is_configured": false, 00:13:21.964 "data_offset": 0, 00:13:21.964 "data_size": 63488 00:13:21.964 } 00:13:21.964 ] 00:13:21.964 }' 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.964 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.535 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.535 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.535 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.535 16:51:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:22.535 16:51:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.535 [2024-09-29 16:51:44.017604] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:22.535 "name": "Existed_Raid", 00:13:22.535 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:22.535 "strip_size_kb": 64, 00:13:22.535 "state": "configuring", 00:13:22.535 "raid_level": "raid5f", 00:13:22.535 "superblock": true, 00:13:22.535 "num_base_bdevs": 3, 00:13:22.535 "num_base_bdevs_discovered": 2, 00:13:22.535 "num_base_bdevs_operational": 3, 00:13:22.535 "base_bdevs_list": [ 00:13:22.535 { 00:13:22.535 "name": "BaseBdev1", 00:13:22.535 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:22.535 "is_configured": true, 00:13:22.535 "data_offset": 2048, 00:13:22.535 "data_size": 63488 00:13:22.535 }, 00:13:22.535 { 00:13:22.535 "name": null, 00:13:22.535 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:22.535 "is_configured": false, 00:13:22.535 "data_offset": 0, 00:13:22.535 "data_size": 63488 00:13:22.535 }, 00:13:22.535 { 00:13:22.535 "name": "BaseBdev3", 00:13:22.535 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:22.535 "is_configured": true, 00:13:22.535 "data_offset": 2048, 00:13:22.535 "data_size": 63488 00:13:22.535 } 00:13:22.535 ] 00:13:22.535 }' 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:22.535 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.105 [2024-09-29 16:51:44.552712] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.105 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.106 "name": "Existed_Raid", 00:13:23.106 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:23.106 "strip_size_kb": 64, 00:13:23.106 "state": "configuring", 00:13:23.106 "raid_level": "raid5f", 00:13:23.106 "superblock": true, 00:13:23.106 "num_base_bdevs": 3, 00:13:23.106 "num_base_bdevs_discovered": 1, 00:13:23.106 "num_base_bdevs_operational": 3, 00:13:23.106 "base_bdevs_list": [ 00:13:23.106 { 00:13:23.106 "name": null, 00:13:23.106 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:23.106 "is_configured": false, 00:13:23.106 "data_offset": 0, 00:13:23.106 "data_size": 63488 00:13:23.106 }, 00:13:23.106 { 00:13:23.106 "name": null, 00:13:23.106 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:23.106 "is_configured": false, 00:13:23.106 "data_offset": 0, 00:13:23.106 "data_size": 63488 00:13:23.106 }, 00:13:23.106 { 00:13:23.106 "name": "BaseBdev3", 00:13:23.106 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:23.106 "is_configured": true, 00:13:23.106 "data_offset": 2048, 00:13:23.106 "data_size": 63488 00:13:23.106 } 00:13:23.106 ] 00:13:23.106 }' 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.106 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.366 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.366 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.366 16:51:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:23.366 16:51:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.366 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.366 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:23.366 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:23.366 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.366 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.366 [2024-09-29 16:51:45.038646] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:23.626 "name": "Existed_Raid", 00:13:23.626 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:23.626 "strip_size_kb": 64, 00:13:23.626 "state": "configuring", 00:13:23.626 "raid_level": "raid5f", 00:13:23.626 "superblock": true, 00:13:23.626 "num_base_bdevs": 3, 00:13:23.626 "num_base_bdevs_discovered": 2, 00:13:23.626 "num_base_bdevs_operational": 3, 00:13:23.626 "base_bdevs_list": [ 00:13:23.626 { 00:13:23.626 "name": null, 00:13:23.626 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:23.626 "is_configured": false, 00:13:23.626 "data_offset": 0, 00:13:23.626 "data_size": 63488 00:13:23.626 }, 00:13:23.626 { 00:13:23.626 "name": "BaseBdev2", 00:13:23.626 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:23.626 "is_configured": true, 00:13:23.626 "data_offset": 2048, 00:13:23.626 "data_size": 63488 00:13:23.626 }, 00:13:23.626 { 00:13:23.626 "name": "BaseBdev3", 00:13:23.626 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:23.626 "is_configured": true, 00:13:23.626 "data_offset": 2048, 00:13:23.626 "data_size": 63488 00:13:23.626 } 00:13:23.626 ] 00:13:23.626 }' 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:23.626 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.886 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.146 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.147 [2024-09-29 16:51:45.612333] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:24.147 [2024-09-29 16:51:45.612593] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:24.147 [2024-09-29 16:51:45.612646] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:24.147 [2024-09-29 16:51:45.612928] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:24.147 NewBaseBdev 00:13:24.147 [2024-09-29 16:51:45.613363] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:24.147 [2024-09-29 16:51:45.613425] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:24.147 [2024-09-29 16:51:45.613610] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.147 [ 00:13:24.147 { 00:13:24.147 "name": "NewBaseBdev", 00:13:24.147 "aliases": [ 00:13:24.147 "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4" 00:13:24.147 ], 00:13:24.147 "product_name": "Malloc disk", 00:13:24.147 "block_size": 512, 00:13:24.147 "num_blocks": 65536, 00:13:24.147 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:24.147 "assigned_rate_limits": { 00:13:24.147 "rw_ios_per_sec": 0, 00:13:24.147 "rw_mbytes_per_sec": 0, 00:13:24.147 "r_mbytes_per_sec": 0, 00:13:24.147 "w_mbytes_per_sec": 0 00:13:24.147 }, 00:13:24.147 "claimed": true, 00:13:24.147 "claim_type": "exclusive_write", 00:13:24.147 "zoned": false, 00:13:24.147 "supported_io_types": { 00:13:24.147 "read": true, 00:13:24.147 "write": true, 00:13:24.147 "unmap": true, 00:13:24.147 "flush": true, 00:13:24.147 "reset": true, 00:13:24.147 "nvme_admin": false, 00:13:24.147 "nvme_io": false, 00:13:24.147 "nvme_io_md": false, 00:13:24.147 "write_zeroes": true, 00:13:24.147 "zcopy": true, 00:13:24.147 "get_zone_info": false, 00:13:24.147 "zone_management": false, 00:13:24.147 "zone_append": false, 00:13:24.147 "compare": false, 00:13:24.147 "compare_and_write": false, 00:13:24.147 "abort": true, 00:13:24.147 "seek_hole": false, 00:13:24.147 "seek_data": false, 00:13:24.147 "copy": true, 00:13:24.147 "nvme_iov_md": false 00:13:24.147 }, 00:13:24.147 "memory_domains": [ 00:13:24.147 { 00:13:24.147 "dma_device_id": "system", 00:13:24.147 "dma_device_type": 1 00:13:24.147 }, 00:13:24.147 { 00:13:24.147 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:24.147 "dma_device_type": 2 00:13:24.147 } 00:13:24.147 ], 00:13:24.147 "driver_specific": {} 00:13:24.147 } 00:13:24.147 ] 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:24.147 "name": "Existed_Raid", 00:13:24.147 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:24.147 "strip_size_kb": 64, 00:13:24.147 "state": "online", 00:13:24.147 "raid_level": "raid5f", 00:13:24.147 "superblock": true, 00:13:24.147 "num_base_bdevs": 3, 00:13:24.147 "num_base_bdevs_discovered": 3, 00:13:24.147 "num_base_bdevs_operational": 3, 00:13:24.147 "base_bdevs_list": [ 00:13:24.147 { 00:13:24.147 "name": "NewBaseBdev", 00:13:24.147 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:24.147 "is_configured": true, 00:13:24.147 "data_offset": 2048, 00:13:24.147 "data_size": 63488 00:13:24.147 }, 00:13:24.147 { 00:13:24.147 "name": "BaseBdev2", 00:13:24.147 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:24.147 "is_configured": true, 00:13:24.147 "data_offset": 2048, 00:13:24.147 "data_size": 63488 00:13:24.147 }, 00:13:24.147 { 00:13:24.147 "name": "BaseBdev3", 00:13:24.147 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:24.147 "is_configured": true, 00:13:24.147 "data_offset": 2048, 00:13:24.147 "data_size": 63488 00:13:24.147 } 00:13:24.147 ] 00:13:24.147 }' 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:24.147 16:51:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.407 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.407 [2024-09-29 16:51:46.071771] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:24.667 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.667 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:24.667 "name": "Existed_Raid", 00:13:24.667 "aliases": [ 00:13:24.667 "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6" 00:13:24.667 ], 00:13:24.667 "product_name": "Raid Volume", 00:13:24.667 "block_size": 512, 00:13:24.667 "num_blocks": 126976, 00:13:24.667 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:24.667 "assigned_rate_limits": { 00:13:24.667 "rw_ios_per_sec": 0, 00:13:24.667 "rw_mbytes_per_sec": 0, 00:13:24.667 "r_mbytes_per_sec": 0, 00:13:24.667 "w_mbytes_per_sec": 0 00:13:24.667 }, 00:13:24.667 "claimed": false, 00:13:24.667 "zoned": false, 00:13:24.667 "supported_io_types": { 00:13:24.667 "read": true, 00:13:24.667 "write": true, 00:13:24.667 "unmap": false, 00:13:24.667 "flush": false, 00:13:24.667 "reset": true, 00:13:24.667 "nvme_admin": false, 00:13:24.667 "nvme_io": false, 00:13:24.667 "nvme_io_md": false, 00:13:24.667 "write_zeroes": true, 00:13:24.667 "zcopy": false, 00:13:24.667 "get_zone_info": false, 00:13:24.667 "zone_management": false, 00:13:24.667 "zone_append": false, 00:13:24.667 "compare": false, 00:13:24.667 "compare_and_write": false, 00:13:24.667 "abort": false, 00:13:24.667 "seek_hole": false, 00:13:24.667 "seek_data": false, 00:13:24.667 "copy": false, 00:13:24.667 "nvme_iov_md": false 00:13:24.667 }, 00:13:24.667 "driver_specific": { 00:13:24.667 "raid": { 00:13:24.667 "uuid": "2c57f1ef-6e82-4ecc-9d1e-763b825ff1f6", 00:13:24.667 "strip_size_kb": 64, 00:13:24.667 "state": "online", 00:13:24.667 "raid_level": "raid5f", 00:13:24.667 "superblock": true, 00:13:24.667 "num_base_bdevs": 3, 00:13:24.667 "num_base_bdevs_discovered": 3, 00:13:24.667 "num_base_bdevs_operational": 3, 00:13:24.667 "base_bdevs_list": [ 00:13:24.667 { 00:13:24.667 "name": "NewBaseBdev", 00:13:24.667 "uuid": "a3c8cd1e-7d58-474c-9237-d47cdb3a4dc4", 00:13:24.667 "is_configured": true, 00:13:24.667 "data_offset": 2048, 00:13:24.667 "data_size": 63488 00:13:24.667 }, 00:13:24.667 { 00:13:24.667 "name": "BaseBdev2", 00:13:24.667 "uuid": "f4c0ac8d-4181-41f7-94e2-993c724ba2d4", 00:13:24.667 "is_configured": true, 00:13:24.667 "data_offset": 2048, 00:13:24.667 "data_size": 63488 00:13:24.667 }, 00:13:24.667 { 00:13:24.667 "name": "BaseBdev3", 00:13:24.667 "uuid": "37a6ef20-9186-4047-b710-af81de7486be", 00:13:24.667 "is_configured": true, 00:13:24.667 "data_offset": 2048, 00:13:24.667 "data_size": 63488 00:13:24.667 } 00:13:24.667 ] 00:13:24.667 } 00:13:24.667 } 00:13:24.668 }' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:24.668 BaseBdev2 00:13:24.668 BaseBdev3' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:24.668 [2024-09-29 16:51:46.323135] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:24.668 [2024-09-29 16:51:46.323201] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:24.668 [2024-09-29 16:51:46.323294] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:24.668 [2024-09-29 16:51:46.323550] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:24.668 [2024-09-29 16:51:46.323606] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 90699 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 90699 ']' 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 90699 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:13:24.668 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:24.927 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90699 00:13:24.927 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:24.927 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:24.927 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90699' 00:13:24.927 killing process with pid 90699 00:13:24.927 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 90699 00:13:24.927 [2024-09-29 16:51:46.372107] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:24.927 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 90699 00:13:24.927 [2024-09-29 16:51:46.403334] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:25.188 16:51:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:13:25.188 00:13:25.188 real 0m9.187s 00:13:25.188 user 0m15.594s 00:13:25.188 sys 0m2.056s 00:13:25.188 ************************************ 00:13:25.188 END TEST raid5f_state_function_test_sb 00:13:25.188 ************************************ 00:13:25.188 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:25.188 16:51:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:25.188 16:51:46 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:13:25.188 16:51:46 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:13:25.188 16:51:46 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:25.188 16:51:46 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:25.188 ************************************ 00:13:25.188 START TEST raid5f_superblock_test 00:13:25.188 ************************************ 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 3 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=91308 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 91308 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 91308 ']' 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:25.188 16:51:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:25.189 16:51:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:25.189 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:25.189 16:51:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:25.189 16:51:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:25.189 [2024-09-29 16:51:46.819104] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:13:25.189 [2024-09-29 16:51:46.819335] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91308 ] 00:13:25.449 [2024-09-29 16:51:46.964909] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:25.449 [2024-09-29 16:51:47.010872] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:25.449 [2024-09-29 16:51:47.053993] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:25.449 [2024-09-29 16:51:47.054031] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:26.018 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.019 malloc1 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.019 [2024-09-29 16:51:47.652899] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:26.019 [2024-09-29 16:51:47.653069] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:26.019 [2024-09-29 16:51:47.653109] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:26.019 [2024-09-29 16:51:47.653166] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:26.019 [2024-09-29 16:51:47.655200] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:26.019 [2024-09-29 16:51:47.655272] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:26.019 pt1 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.019 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.279 malloc2 00:13:26.279 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.279 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:26.279 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.280 [2024-09-29 16:51:47.706306] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:26.280 [2024-09-29 16:51:47.706545] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:26.280 [2024-09-29 16:51:47.706638] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:26.280 [2024-09-29 16:51:47.706675] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:26.280 [2024-09-29 16:51:47.711266] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:26.280 [2024-09-29 16:51:47.711329] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:26.280 pt2 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.280 malloc3 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.280 [2024-09-29 16:51:47.737688] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:26.280 [2024-09-29 16:51:47.737832] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:26.280 [2024-09-29 16:51:47.737865] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:26.280 [2024-09-29 16:51:47.737894] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:26.280 [2024-09-29 16:51:47.739917] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:26.280 [2024-09-29 16:51:47.740011] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:26.280 pt3 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.280 [2024-09-29 16:51:47.749769] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:26.280 [2024-09-29 16:51:47.751684] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:26.280 [2024-09-29 16:51:47.751786] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:26.280 [2024-09-29 16:51:47.751993] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:26.280 [2024-09-29 16:51:47.752057] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:26.280 [2024-09-29 16:51:47.752339] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:26.280 [2024-09-29 16:51:47.752830] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:26.280 [2024-09-29 16:51:47.752883] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:26.280 [2024-09-29 16:51:47.753071] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:26.280 "name": "raid_bdev1", 00:13:26.280 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:26.280 "strip_size_kb": 64, 00:13:26.280 "state": "online", 00:13:26.280 "raid_level": "raid5f", 00:13:26.280 "superblock": true, 00:13:26.280 "num_base_bdevs": 3, 00:13:26.280 "num_base_bdevs_discovered": 3, 00:13:26.280 "num_base_bdevs_operational": 3, 00:13:26.280 "base_bdevs_list": [ 00:13:26.280 { 00:13:26.280 "name": "pt1", 00:13:26.280 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:26.280 "is_configured": true, 00:13:26.280 "data_offset": 2048, 00:13:26.280 "data_size": 63488 00:13:26.280 }, 00:13:26.280 { 00:13:26.280 "name": "pt2", 00:13:26.280 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:26.280 "is_configured": true, 00:13:26.280 "data_offset": 2048, 00:13:26.280 "data_size": 63488 00:13:26.280 }, 00:13:26.280 { 00:13:26.280 "name": "pt3", 00:13:26.280 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:26.280 "is_configured": true, 00:13:26.280 "data_offset": 2048, 00:13:26.280 "data_size": 63488 00:13:26.280 } 00:13:26.280 ] 00:13:26.280 }' 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:26.280 16:51:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.540 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:13:26.540 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:26.540 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:26.540 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:26.540 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:26.540 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:26.540 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:26.541 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.541 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.541 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:26.541 [2024-09-29 16:51:48.197862] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:26.801 "name": "raid_bdev1", 00:13:26.801 "aliases": [ 00:13:26.801 "a5f1b08d-bed9-4fa6-be03-182466acc1bb" 00:13:26.801 ], 00:13:26.801 "product_name": "Raid Volume", 00:13:26.801 "block_size": 512, 00:13:26.801 "num_blocks": 126976, 00:13:26.801 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:26.801 "assigned_rate_limits": { 00:13:26.801 "rw_ios_per_sec": 0, 00:13:26.801 "rw_mbytes_per_sec": 0, 00:13:26.801 "r_mbytes_per_sec": 0, 00:13:26.801 "w_mbytes_per_sec": 0 00:13:26.801 }, 00:13:26.801 "claimed": false, 00:13:26.801 "zoned": false, 00:13:26.801 "supported_io_types": { 00:13:26.801 "read": true, 00:13:26.801 "write": true, 00:13:26.801 "unmap": false, 00:13:26.801 "flush": false, 00:13:26.801 "reset": true, 00:13:26.801 "nvme_admin": false, 00:13:26.801 "nvme_io": false, 00:13:26.801 "nvme_io_md": false, 00:13:26.801 "write_zeroes": true, 00:13:26.801 "zcopy": false, 00:13:26.801 "get_zone_info": false, 00:13:26.801 "zone_management": false, 00:13:26.801 "zone_append": false, 00:13:26.801 "compare": false, 00:13:26.801 "compare_and_write": false, 00:13:26.801 "abort": false, 00:13:26.801 "seek_hole": false, 00:13:26.801 "seek_data": false, 00:13:26.801 "copy": false, 00:13:26.801 "nvme_iov_md": false 00:13:26.801 }, 00:13:26.801 "driver_specific": { 00:13:26.801 "raid": { 00:13:26.801 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:26.801 "strip_size_kb": 64, 00:13:26.801 "state": "online", 00:13:26.801 "raid_level": "raid5f", 00:13:26.801 "superblock": true, 00:13:26.801 "num_base_bdevs": 3, 00:13:26.801 "num_base_bdevs_discovered": 3, 00:13:26.801 "num_base_bdevs_operational": 3, 00:13:26.801 "base_bdevs_list": [ 00:13:26.801 { 00:13:26.801 "name": "pt1", 00:13:26.801 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:26.801 "is_configured": true, 00:13:26.801 "data_offset": 2048, 00:13:26.801 "data_size": 63488 00:13:26.801 }, 00:13:26.801 { 00:13:26.801 "name": "pt2", 00:13:26.801 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:26.801 "is_configured": true, 00:13:26.801 "data_offset": 2048, 00:13:26.801 "data_size": 63488 00:13:26.801 }, 00:13:26.801 { 00:13:26.801 "name": "pt3", 00:13:26.801 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:26.801 "is_configured": true, 00:13:26.801 "data_offset": 2048, 00:13:26.801 "data_size": 63488 00:13:26.801 } 00:13:26.801 ] 00:13:26.801 } 00:13:26.801 } 00:13:26.801 }' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:26.801 pt2 00:13:26.801 pt3' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:26.801 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 [2024-09-29 16:51:48.481299] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a5f1b08d-bed9-4fa6-be03-182466acc1bb 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a5f1b08d-bed9-4fa6-be03-182466acc1bb ']' 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 [2024-09-29 16:51:48.525071] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:27.063 [2024-09-29 16:51:48.525141] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:27.063 [2024-09-29 16:51:48.525241] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:27.063 [2024-09-29 16:51:48.525351] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:27.063 [2024-09-29 16:51:48.525424] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 [2024-09-29 16:51:48.680859] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:13:27.063 [2024-09-29 16:51:48.682782] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:13:27.063 [2024-09-29 16:51:48.682860] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:13:27.063 [2024-09-29 16:51:48.682924] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:13:27.063 [2024-09-29 16:51:48.683034] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:13:27.063 [2024-09-29 16:51:48.683105] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:13:27.063 [2024-09-29 16:51:48.683151] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:27.063 [2024-09-29 16:51:48.683192] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:13:27.063 request: 00:13:27.063 { 00:13:27.063 "name": "raid_bdev1", 00:13:27.063 "raid_level": "raid5f", 00:13:27.063 "base_bdevs": [ 00:13:27.063 "malloc1", 00:13:27.063 "malloc2", 00:13:27.063 "malloc3" 00:13:27.063 ], 00:13:27.063 "strip_size_kb": 64, 00:13:27.063 "superblock": false, 00:13:27.063 "method": "bdev_raid_create", 00:13:27.063 "req_id": 1 00:13:27.063 } 00:13:27.063 Got JSON-RPC error response 00:13:27.063 response: 00:13:27.063 { 00:13:27.063 "code": -17, 00:13:27.063 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:13:27.063 } 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.063 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.324 [2024-09-29 16:51:48.748698] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:27.324 [2024-09-29 16:51:48.748818] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.324 [2024-09-29 16:51:48.748837] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:27.324 [2024-09-29 16:51:48.748848] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.324 [2024-09-29 16:51:48.750962] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.324 [2024-09-29 16:51:48.751001] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:27.324 [2024-09-29 16:51:48.751074] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:27.324 [2024-09-29 16:51:48.751107] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:27.324 pt1 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.324 "name": "raid_bdev1", 00:13:27.324 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:27.324 "strip_size_kb": 64, 00:13:27.324 "state": "configuring", 00:13:27.324 "raid_level": "raid5f", 00:13:27.324 "superblock": true, 00:13:27.324 "num_base_bdevs": 3, 00:13:27.324 "num_base_bdevs_discovered": 1, 00:13:27.324 "num_base_bdevs_operational": 3, 00:13:27.324 "base_bdevs_list": [ 00:13:27.324 { 00:13:27.324 "name": "pt1", 00:13:27.324 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:27.324 "is_configured": true, 00:13:27.324 "data_offset": 2048, 00:13:27.324 "data_size": 63488 00:13:27.324 }, 00:13:27.324 { 00:13:27.324 "name": null, 00:13:27.324 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:27.324 "is_configured": false, 00:13:27.324 "data_offset": 2048, 00:13:27.324 "data_size": 63488 00:13:27.324 }, 00:13:27.324 { 00:13:27.324 "name": null, 00:13:27.324 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:27.324 "is_configured": false, 00:13:27.324 "data_offset": 2048, 00:13:27.324 "data_size": 63488 00:13:27.324 } 00:13:27.324 ] 00:13:27.324 }' 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.324 16:51:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.584 [2024-09-29 16:51:49.167968] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:27.584 [2024-09-29 16:51:49.168091] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:27.584 [2024-09-29 16:51:49.168128] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:27.584 [2024-09-29 16:51:49.168161] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:27.584 [2024-09-29 16:51:49.168538] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:27.584 [2024-09-29 16:51:49.168601] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:27.584 [2024-09-29 16:51:49.168698] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:27.584 [2024-09-29 16:51:49.168761] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:27.584 pt2 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.584 [2024-09-29 16:51:49.179953] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.584 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.585 "name": "raid_bdev1", 00:13:27.585 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:27.585 "strip_size_kb": 64, 00:13:27.585 "state": "configuring", 00:13:27.585 "raid_level": "raid5f", 00:13:27.585 "superblock": true, 00:13:27.585 "num_base_bdevs": 3, 00:13:27.585 "num_base_bdevs_discovered": 1, 00:13:27.585 "num_base_bdevs_operational": 3, 00:13:27.585 "base_bdevs_list": [ 00:13:27.585 { 00:13:27.585 "name": "pt1", 00:13:27.585 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:27.585 "is_configured": true, 00:13:27.585 "data_offset": 2048, 00:13:27.585 "data_size": 63488 00:13:27.585 }, 00:13:27.585 { 00:13:27.585 "name": null, 00:13:27.585 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:27.585 "is_configured": false, 00:13:27.585 "data_offset": 0, 00:13:27.585 "data_size": 63488 00:13:27.585 }, 00:13:27.585 { 00:13:27.585 "name": null, 00:13:27.585 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:27.585 "is_configured": false, 00:13:27.585 "data_offset": 2048, 00:13:27.585 "data_size": 63488 00:13:27.585 } 00:13:27.585 ] 00:13:27.585 }' 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.585 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.155 [2024-09-29 16:51:49.635152] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:28.155 [2024-09-29 16:51:49.635252] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:28.155 [2024-09-29 16:51:49.635284] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:13:28.155 [2024-09-29 16:51:49.635310] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:28.155 [2024-09-29 16:51:49.635643] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:28.155 [2024-09-29 16:51:49.635697] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:28.155 [2024-09-29 16:51:49.635791] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:28.155 [2024-09-29 16:51:49.635837] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:28.155 pt2 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.155 [2024-09-29 16:51:49.647138] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:28.155 [2024-09-29 16:51:49.647234] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:28.155 [2024-09-29 16:51:49.647276] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:13:28.155 [2024-09-29 16:51:49.647301] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:28.155 [2024-09-29 16:51:49.647635] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:28.155 [2024-09-29 16:51:49.647688] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:28.155 [2024-09-29 16:51:49.647785] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:28.155 [2024-09-29 16:51:49.647846] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:28.155 [2024-09-29 16:51:49.647977] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:28.155 [2024-09-29 16:51:49.648013] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:28.155 [2024-09-29 16:51:49.648256] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:28.155 [2024-09-29 16:51:49.648650] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:28.155 [2024-09-29 16:51:49.648698] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:13:28.155 [2024-09-29 16:51:49.648861] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:28.155 pt3 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.155 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.155 "name": "raid_bdev1", 00:13:28.155 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:28.155 "strip_size_kb": 64, 00:13:28.155 "state": "online", 00:13:28.155 "raid_level": "raid5f", 00:13:28.155 "superblock": true, 00:13:28.155 "num_base_bdevs": 3, 00:13:28.155 "num_base_bdevs_discovered": 3, 00:13:28.155 "num_base_bdevs_operational": 3, 00:13:28.155 "base_bdevs_list": [ 00:13:28.155 { 00:13:28.155 "name": "pt1", 00:13:28.155 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:28.155 "is_configured": true, 00:13:28.155 "data_offset": 2048, 00:13:28.155 "data_size": 63488 00:13:28.155 }, 00:13:28.155 { 00:13:28.155 "name": "pt2", 00:13:28.155 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:28.155 "is_configured": true, 00:13:28.155 "data_offset": 2048, 00:13:28.155 "data_size": 63488 00:13:28.155 }, 00:13:28.155 { 00:13:28.155 "name": "pt3", 00:13:28.155 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:28.155 "is_configured": true, 00:13:28.156 "data_offset": 2048, 00:13:28.156 "data_size": 63488 00:13:28.156 } 00:13:28.156 ] 00:13:28.156 }' 00:13:28.156 16:51:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.156 16:51:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.726 [2024-09-29 16:51:50.146537] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:28.726 "name": "raid_bdev1", 00:13:28.726 "aliases": [ 00:13:28.726 "a5f1b08d-bed9-4fa6-be03-182466acc1bb" 00:13:28.726 ], 00:13:28.726 "product_name": "Raid Volume", 00:13:28.726 "block_size": 512, 00:13:28.726 "num_blocks": 126976, 00:13:28.726 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:28.726 "assigned_rate_limits": { 00:13:28.726 "rw_ios_per_sec": 0, 00:13:28.726 "rw_mbytes_per_sec": 0, 00:13:28.726 "r_mbytes_per_sec": 0, 00:13:28.726 "w_mbytes_per_sec": 0 00:13:28.726 }, 00:13:28.726 "claimed": false, 00:13:28.726 "zoned": false, 00:13:28.726 "supported_io_types": { 00:13:28.726 "read": true, 00:13:28.726 "write": true, 00:13:28.726 "unmap": false, 00:13:28.726 "flush": false, 00:13:28.726 "reset": true, 00:13:28.726 "nvme_admin": false, 00:13:28.726 "nvme_io": false, 00:13:28.726 "nvme_io_md": false, 00:13:28.726 "write_zeroes": true, 00:13:28.726 "zcopy": false, 00:13:28.726 "get_zone_info": false, 00:13:28.726 "zone_management": false, 00:13:28.726 "zone_append": false, 00:13:28.726 "compare": false, 00:13:28.726 "compare_and_write": false, 00:13:28.726 "abort": false, 00:13:28.726 "seek_hole": false, 00:13:28.726 "seek_data": false, 00:13:28.726 "copy": false, 00:13:28.726 "nvme_iov_md": false 00:13:28.726 }, 00:13:28.726 "driver_specific": { 00:13:28.726 "raid": { 00:13:28.726 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:28.726 "strip_size_kb": 64, 00:13:28.726 "state": "online", 00:13:28.726 "raid_level": "raid5f", 00:13:28.726 "superblock": true, 00:13:28.726 "num_base_bdevs": 3, 00:13:28.726 "num_base_bdevs_discovered": 3, 00:13:28.726 "num_base_bdevs_operational": 3, 00:13:28.726 "base_bdevs_list": [ 00:13:28.726 { 00:13:28.726 "name": "pt1", 00:13:28.726 "uuid": "00000000-0000-0000-0000-000000000001", 00:13:28.726 "is_configured": true, 00:13:28.726 "data_offset": 2048, 00:13:28.726 "data_size": 63488 00:13:28.726 }, 00:13:28.726 { 00:13:28.726 "name": "pt2", 00:13:28.726 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:28.726 "is_configured": true, 00:13:28.726 "data_offset": 2048, 00:13:28.726 "data_size": 63488 00:13:28.726 }, 00:13:28.726 { 00:13:28.726 "name": "pt3", 00:13:28.726 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:28.726 "is_configured": true, 00:13:28.726 "data_offset": 2048, 00:13:28.726 "data_size": 63488 00:13:28.726 } 00:13:28.726 ] 00:13:28.726 } 00:13:28.726 } 00:13:28.726 }' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:13:28.726 pt2 00:13:28.726 pt3' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.726 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.987 [2024-09-29 16:51:50.433981] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a5f1b08d-bed9-4fa6-be03-182466acc1bb '!=' a5f1b08d-bed9-4fa6-be03-182466acc1bb ']' 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.987 [2024-09-29 16:51:50.477841] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:28.987 "name": "raid_bdev1", 00:13:28.987 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:28.987 "strip_size_kb": 64, 00:13:28.987 "state": "online", 00:13:28.987 "raid_level": "raid5f", 00:13:28.987 "superblock": true, 00:13:28.987 "num_base_bdevs": 3, 00:13:28.987 "num_base_bdevs_discovered": 2, 00:13:28.987 "num_base_bdevs_operational": 2, 00:13:28.987 "base_bdevs_list": [ 00:13:28.987 { 00:13:28.987 "name": null, 00:13:28.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:28.987 "is_configured": false, 00:13:28.987 "data_offset": 0, 00:13:28.987 "data_size": 63488 00:13:28.987 }, 00:13:28.987 { 00:13:28.987 "name": "pt2", 00:13:28.987 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:28.987 "is_configured": true, 00:13:28.987 "data_offset": 2048, 00:13:28.987 "data_size": 63488 00:13:28.987 }, 00:13:28.987 { 00:13:28.987 "name": "pt3", 00:13:28.987 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:28.987 "is_configured": true, 00:13:28.987 "data_offset": 2048, 00:13:28.987 "data_size": 63488 00:13:28.987 } 00:13:28.987 ] 00:13:28.987 }' 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:28.987 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.557 [2024-09-29 16:51:50.945005] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:29.557 [2024-09-29 16:51:50.945082] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:29.557 [2024-09-29 16:51:50.945179] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:29.557 [2024-09-29 16:51:50.945289] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:29.557 [2024-09-29 16:51:50.945339] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:13:29.557 16:51:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.557 [2024-09-29 16:51:51.028842] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:13:29.557 [2024-09-29 16:51:51.028957] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:29.557 [2024-09-29 16:51:51.028995] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:13:29.557 [2024-09-29 16:51:51.029023] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:29.557 [2024-09-29 16:51:51.031146] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:29.557 [2024-09-29 16:51:51.031217] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:13:29.557 [2024-09-29 16:51:51.031301] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:13:29.557 [2024-09-29 16:51:51.031377] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:29.557 pt2 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:29.557 "name": "raid_bdev1", 00:13:29.557 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:29.557 "strip_size_kb": 64, 00:13:29.557 "state": "configuring", 00:13:29.557 "raid_level": "raid5f", 00:13:29.557 "superblock": true, 00:13:29.557 "num_base_bdevs": 3, 00:13:29.557 "num_base_bdevs_discovered": 1, 00:13:29.557 "num_base_bdevs_operational": 2, 00:13:29.557 "base_bdevs_list": [ 00:13:29.557 { 00:13:29.557 "name": null, 00:13:29.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:29.557 "is_configured": false, 00:13:29.557 "data_offset": 2048, 00:13:29.557 "data_size": 63488 00:13:29.557 }, 00:13:29.557 { 00:13:29.557 "name": "pt2", 00:13:29.557 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:29.557 "is_configured": true, 00:13:29.557 "data_offset": 2048, 00:13:29.557 "data_size": 63488 00:13:29.557 }, 00:13:29.557 { 00:13:29.557 "name": null, 00:13:29.557 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:29.557 "is_configured": false, 00:13:29.557 "data_offset": 2048, 00:13:29.557 "data_size": 63488 00:13:29.557 } 00:13:29.557 ] 00:13:29.557 }' 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:29.557 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:29.818 [2024-09-29 16:51:51.472096] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:29.818 [2024-09-29 16:51:51.472190] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:29.818 [2024-09-29 16:51:51.472223] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:13:29.818 [2024-09-29 16:51:51.472249] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:29.818 [2024-09-29 16:51:51.472632] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:29.818 [2024-09-29 16:51:51.472689] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:29.818 [2024-09-29 16:51:51.472794] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:29.818 [2024-09-29 16:51:51.472843] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:29.818 [2024-09-29 16:51:51.472970] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:29.818 [2024-09-29 16:51:51.473006] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:29.818 [2024-09-29 16:51:51.473260] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:29.818 [2024-09-29 16:51:51.473761] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:29.818 [2024-09-29 16:51:51.473829] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:13:29.818 [2024-09-29 16:51:51.474101] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:29.818 pt3 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:29.818 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.078 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.078 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.078 "name": "raid_bdev1", 00:13:30.078 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:30.078 "strip_size_kb": 64, 00:13:30.078 "state": "online", 00:13:30.078 "raid_level": "raid5f", 00:13:30.078 "superblock": true, 00:13:30.078 "num_base_bdevs": 3, 00:13:30.078 "num_base_bdevs_discovered": 2, 00:13:30.078 "num_base_bdevs_operational": 2, 00:13:30.078 "base_bdevs_list": [ 00:13:30.078 { 00:13:30.078 "name": null, 00:13:30.078 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.078 "is_configured": false, 00:13:30.078 "data_offset": 2048, 00:13:30.078 "data_size": 63488 00:13:30.078 }, 00:13:30.078 { 00:13:30.078 "name": "pt2", 00:13:30.078 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:30.078 "is_configured": true, 00:13:30.078 "data_offset": 2048, 00:13:30.078 "data_size": 63488 00:13:30.078 }, 00:13:30.078 { 00:13:30.078 "name": "pt3", 00:13:30.078 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:30.078 "is_configured": true, 00:13:30.078 "data_offset": 2048, 00:13:30.078 "data_size": 63488 00:13:30.078 } 00:13:30.078 ] 00:13:30.078 }' 00:13:30.078 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.078 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.338 [2024-09-29 16:51:51.907356] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:30.338 [2024-09-29 16:51:51.907427] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:30.338 [2024-09-29 16:51:51.907495] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:30.338 [2024-09-29 16:51:51.907574] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:30.338 [2024-09-29 16:51:51.907640] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:13:30.338 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.339 [2024-09-29 16:51:51.971247] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:13:30.339 [2024-09-29 16:51:51.971348] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.339 [2024-09-29 16:51:51.971379] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:30.339 [2024-09-29 16:51:51.971409] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.339 [2024-09-29 16:51:51.973599] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.339 [2024-09-29 16:51:51.973690] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:13:30.339 [2024-09-29 16:51:51.973783] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:13:30.339 [2024-09-29 16:51:51.973855] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:13:30.339 [2024-09-29 16:51:51.974012] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:13:30.339 [2024-09-29 16:51:51.974075] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:30.339 [2024-09-29 16:51:51.974128] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:13:30.339 [2024-09-29 16:51:51.974243] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:13:30.339 pt1 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.339 16:51:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.599 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.599 "name": "raid_bdev1", 00:13:30.599 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:30.599 "strip_size_kb": 64, 00:13:30.599 "state": "configuring", 00:13:30.599 "raid_level": "raid5f", 00:13:30.599 "superblock": true, 00:13:30.599 "num_base_bdevs": 3, 00:13:30.599 "num_base_bdevs_discovered": 1, 00:13:30.599 "num_base_bdevs_operational": 2, 00:13:30.599 "base_bdevs_list": [ 00:13:30.599 { 00:13:30.599 "name": null, 00:13:30.599 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:30.599 "is_configured": false, 00:13:30.599 "data_offset": 2048, 00:13:30.599 "data_size": 63488 00:13:30.599 }, 00:13:30.599 { 00:13:30.599 "name": "pt2", 00:13:30.599 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:30.599 "is_configured": true, 00:13:30.599 "data_offset": 2048, 00:13:30.599 "data_size": 63488 00:13:30.599 }, 00:13:30.599 { 00:13:30.599 "name": null, 00:13:30.599 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:30.599 "is_configured": false, 00:13:30.599 "data_offset": 2048, 00:13:30.599 "data_size": 63488 00:13:30.599 } 00:13:30.599 ] 00:13:30.599 }' 00:13:30.599 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.599 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.859 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.859 [2024-09-29 16:51:52.482489] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:13:30.859 [2024-09-29 16:51:52.482590] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.859 [2024-09-29 16:51:52.482623] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:13:30.859 [2024-09-29 16:51:52.482653] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.859 [2024-09-29 16:51:52.483093] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.859 [2024-09-29 16:51:52.483159] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:13:30.859 [2024-09-29 16:51:52.483266] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:13:30.859 [2024-09-29 16:51:52.483335] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:13:30.859 [2024-09-29 16:51:52.483462] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:13:30.860 [2024-09-29 16:51:52.483503] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:30.860 [2024-09-29 16:51:52.483776] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:30.860 [2024-09-29 16:51:52.484245] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:13:30.860 [2024-09-29 16:51:52.484294] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:13:30.860 [2024-09-29 16:51:52.484513] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:30.860 pt3 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:30.860 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.119 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:31.119 "name": "raid_bdev1", 00:13:31.119 "uuid": "a5f1b08d-bed9-4fa6-be03-182466acc1bb", 00:13:31.119 "strip_size_kb": 64, 00:13:31.119 "state": "online", 00:13:31.119 "raid_level": "raid5f", 00:13:31.119 "superblock": true, 00:13:31.119 "num_base_bdevs": 3, 00:13:31.120 "num_base_bdevs_discovered": 2, 00:13:31.120 "num_base_bdevs_operational": 2, 00:13:31.120 "base_bdevs_list": [ 00:13:31.120 { 00:13:31.120 "name": null, 00:13:31.120 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:31.120 "is_configured": false, 00:13:31.120 "data_offset": 2048, 00:13:31.120 "data_size": 63488 00:13:31.120 }, 00:13:31.120 { 00:13:31.120 "name": "pt2", 00:13:31.120 "uuid": "00000000-0000-0000-0000-000000000002", 00:13:31.120 "is_configured": true, 00:13:31.120 "data_offset": 2048, 00:13:31.120 "data_size": 63488 00:13:31.120 }, 00:13:31.120 { 00:13:31.120 "name": "pt3", 00:13:31.120 "uuid": "00000000-0000-0000-0000-000000000003", 00:13:31.120 "is_configured": true, 00:13:31.120 "data_offset": 2048, 00:13:31.120 "data_size": 63488 00:13:31.120 } 00:13:31.120 ] 00:13:31.120 }' 00:13:31.120 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:31.120 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.380 16:51:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.380 [2024-09-29 16:51:52.993811] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' a5f1b08d-bed9-4fa6-be03-182466acc1bb '!=' a5f1b08d-bed9-4fa6-be03-182466acc1bb ']' 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 91308 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 91308 ']' 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 91308 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:31.380 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91308 00:13:31.640 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:31.640 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:31.640 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91308' 00:13:31.640 killing process with pid 91308 00:13:31.640 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 91308 00:13:31.640 [2024-09-29 16:51:53.069529] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:31.640 [2024-09-29 16:51:53.069671] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:31.640 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 91308 00:13:31.640 [2024-09-29 16:51:53.069749] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:31.640 [2024-09-29 16:51:53.069761] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:13:31.640 [2024-09-29 16:51:53.104101] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:31.901 16:51:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:13:31.901 00:13:31.901 real 0m6.621s 00:13:31.901 user 0m11.075s 00:13:31.901 sys 0m1.434s 00:13:31.901 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:31.901 ************************************ 00:13:31.901 END TEST raid5f_superblock_test 00:13:31.901 ************************************ 00:13:31.901 16:51:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.901 16:51:53 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:13:31.901 16:51:53 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:13:31.901 16:51:53 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:13:31.901 16:51:53 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:31.901 16:51:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:31.901 ************************************ 00:13:31.901 START TEST raid5f_rebuild_test 00:13:31.901 ************************************ 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 false false true 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=91735 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 91735 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 91735 ']' 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:31.901 16:51:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:31.901 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:31.902 16:51:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:31.902 16:51:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:31.902 16:51:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:31.902 [2024-09-29 16:51:53.534691] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:13:31.902 [2024-09-29 16:51:53.534928] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:13:31.902 Zero copy mechanism will not be used. 00:13:31.902 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91735 ] 00:13:32.161 [2024-09-29 16:51:53.659968] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:32.161 [2024-09-29 16:51:53.703744] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:32.161 [2024-09-29 16:51:53.746496] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:32.161 [2024-09-29 16:51:53.746612] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.732 BaseBdev1_malloc 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.732 [2024-09-29 16:51:54.369269] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:32.732 [2024-09-29 16:51:54.369340] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.732 [2024-09-29 16:51:54.369368] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:32.732 [2024-09-29 16:51:54.369381] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.732 [2024-09-29 16:51:54.371451] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.732 [2024-09-29 16:51:54.371564] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:32.732 BaseBdev1 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.732 BaseBdev2_malloc 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.732 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.992 [2024-09-29 16:51:54.407799] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:32.992 [2024-09-29 16:51:54.407931] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.992 [2024-09-29 16:51:54.407977] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:32.992 [2024-09-29 16:51:54.408010] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.992 [2024-09-29 16:51:54.410102] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.992 [2024-09-29 16:51:54.410186] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:32.992 BaseBdev2 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.992 BaseBdev3_malloc 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.992 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.992 [2024-09-29 16:51:54.436252] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:32.992 [2024-09-29 16:51:54.436363] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.992 [2024-09-29 16:51:54.436396] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:32.993 [2024-09-29 16:51:54.436404] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.993 [2024-09-29 16:51:54.438367] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.993 [2024-09-29 16:51:54.438404] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:32.993 BaseBdev3 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.993 spare_malloc 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.993 spare_delay 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.993 [2024-09-29 16:51:54.476828] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:32.993 [2024-09-29 16:51:54.476881] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:32.993 [2024-09-29 16:51:54.476908] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:32.993 [2024-09-29 16:51:54.476917] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:32.993 [2024-09-29 16:51:54.478950] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:32.993 [2024-09-29 16:51:54.479063] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:32.993 spare 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.993 [2024-09-29 16:51:54.488879] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:32.993 [2024-09-29 16:51:54.490682] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:32.993 [2024-09-29 16:51:54.490810] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:32.993 [2024-09-29 16:51:54.490904] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:32.993 [2024-09-29 16:51:54.490961] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:32.993 [2024-09-29 16:51:54.491278] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:32.993 [2024-09-29 16:51:54.491760] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:32.993 [2024-09-29 16:51:54.491808] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:32.993 [2024-09-29 16:51:54.492003] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.993 "name": "raid_bdev1", 00:13:32.993 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:32.993 "strip_size_kb": 64, 00:13:32.993 "state": "online", 00:13:32.993 "raid_level": "raid5f", 00:13:32.993 "superblock": false, 00:13:32.993 "num_base_bdevs": 3, 00:13:32.993 "num_base_bdevs_discovered": 3, 00:13:32.993 "num_base_bdevs_operational": 3, 00:13:32.993 "base_bdevs_list": [ 00:13:32.993 { 00:13:32.993 "name": "BaseBdev1", 00:13:32.993 "uuid": "92688e0b-4ad4-5d7c-ac88-33325c07d1fd", 00:13:32.993 "is_configured": true, 00:13:32.993 "data_offset": 0, 00:13:32.993 "data_size": 65536 00:13:32.993 }, 00:13:32.993 { 00:13:32.993 "name": "BaseBdev2", 00:13:32.993 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:32.993 "is_configured": true, 00:13:32.993 "data_offset": 0, 00:13:32.993 "data_size": 65536 00:13:32.993 }, 00:13:32.993 { 00:13:32.993 "name": "BaseBdev3", 00:13:32.993 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:32.993 "is_configured": true, 00:13:32.993 "data_offset": 0, 00:13:32.993 "data_size": 65536 00:13:32.993 } 00:13:32.993 ] 00:13:32.993 }' 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.993 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.563 [2024-09-29 16:51:54.956752] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.563 16:51:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:33.564 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:33.564 [2024-09-29 16:51:55.224190] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:33.824 /dev/nbd0 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:33.824 1+0 records in 00:13:33.824 1+0 records out 00:13:33.824 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000387223 s, 10.6 MB/s 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:33.824 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:13:34.086 512+0 records in 00:13:34.086 512+0 records out 00:13:34.086 67108864 bytes (67 MB, 64 MiB) copied, 0.292053 s, 230 MB/s 00:13:34.086 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:34.086 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:34.086 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:34.086 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:34.086 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:34.086 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:34.086 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:34.346 [2024-09-29 16:51:55.814849] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.346 [2024-09-29 16:51:55.830904] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:34.346 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:34.347 "name": "raid_bdev1", 00:13:34.347 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:34.347 "strip_size_kb": 64, 00:13:34.347 "state": "online", 00:13:34.347 "raid_level": "raid5f", 00:13:34.347 "superblock": false, 00:13:34.347 "num_base_bdevs": 3, 00:13:34.347 "num_base_bdevs_discovered": 2, 00:13:34.347 "num_base_bdevs_operational": 2, 00:13:34.347 "base_bdevs_list": [ 00:13:34.347 { 00:13:34.347 "name": null, 00:13:34.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:34.347 "is_configured": false, 00:13:34.347 "data_offset": 0, 00:13:34.347 "data_size": 65536 00:13:34.347 }, 00:13:34.347 { 00:13:34.347 "name": "BaseBdev2", 00:13:34.347 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:34.347 "is_configured": true, 00:13:34.347 "data_offset": 0, 00:13:34.347 "data_size": 65536 00:13:34.347 }, 00:13:34.347 { 00:13:34.347 "name": "BaseBdev3", 00:13:34.347 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:34.347 "is_configured": true, 00:13:34.347 "data_offset": 0, 00:13:34.347 "data_size": 65536 00:13:34.347 } 00:13:34.347 ] 00:13:34.347 }' 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:34.347 16:51:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.607 16:51:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:34.607 16:51:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.607 16:51:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:34.607 [2024-09-29 16:51:56.278137] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:34.867 [2024-09-29 16:51:56.281924] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027cd0 00:13:34.867 [2024-09-29 16:51:56.284140] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:34.867 16:51:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:34.867 16:51:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.805 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:35.805 "name": "raid_bdev1", 00:13:35.805 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:35.805 "strip_size_kb": 64, 00:13:35.805 "state": "online", 00:13:35.805 "raid_level": "raid5f", 00:13:35.805 "superblock": false, 00:13:35.805 "num_base_bdevs": 3, 00:13:35.805 "num_base_bdevs_discovered": 3, 00:13:35.805 "num_base_bdevs_operational": 3, 00:13:35.805 "process": { 00:13:35.805 "type": "rebuild", 00:13:35.805 "target": "spare", 00:13:35.805 "progress": { 00:13:35.805 "blocks": 20480, 00:13:35.805 "percent": 15 00:13:35.805 } 00:13:35.805 }, 00:13:35.805 "base_bdevs_list": [ 00:13:35.805 { 00:13:35.805 "name": "spare", 00:13:35.805 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:35.805 "is_configured": true, 00:13:35.805 "data_offset": 0, 00:13:35.805 "data_size": 65536 00:13:35.805 }, 00:13:35.805 { 00:13:35.805 "name": "BaseBdev2", 00:13:35.805 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:35.805 "is_configured": true, 00:13:35.805 "data_offset": 0, 00:13:35.805 "data_size": 65536 00:13:35.805 }, 00:13:35.805 { 00:13:35.805 "name": "BaseBdev3", 00:13:35.805 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:35.805 "is_configured": true, 00:13:35.805 "data_offset": 0, 00:13:35.805 "data_size": 65536 00:13:35.806 } 00:13:35.806 ] 00:13:35.806 }' 00:13:35.806 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:35.806 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:35.806 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:35.806 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:35.806 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:35.806 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.806 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:35.806 [2024-09-29 16:51:57.455157] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:36.065 [2024-09-29 16:51:57.491130] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:36.065 [2024-09-29 16:51:57.491185] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:36.065 [2024-09-29 16:51:57.491201] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:36.065 [2024-09-29 16:51:57.491210] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:36.065 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.065 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:36.065 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:36.065 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:36.065 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:36.065 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:36.066 "name": "raid_bdev1", 00:13:36.066 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:36.066 "strip_size_kb": 64, 00:13:36.066 "state": "online", 00:13:36.066 "raid_level": "raid5f", 00:13:36.066 "superblock": false, 00:13:36.066 "num_base_bdevs": 3, 00:13:36.066 "num_base_bdevs_discovered": 2, 00:13:36.066 "num_base_bdevs_operational": 2, 00:13:36.066 "base_bdevs_list": [ 00:13:36.066 { 00:13:36.066 "name": null, 00:13:36.066 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.066 "is_configured": false, 00:13:36.066 "data_offset": 0, 00:13:36.066 "data_size": 65536 00:13:36.066 }, 00:13:36.066 { 00:13:36.066 "name": "BaseBdev2", 00:13:36.066 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:36.066 "is_configured": true, 00:13:36.066 "data_offset": 0, 00:13:36.066 "data_size": 65536 00:13:36.066 }, 00:13:36.066 { 00:13:36.066 "name": "BaseBdev3", 00:13:36.066 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:36.066 "is_configured": true, 00:13:36.066 "data_offset": 0, 00:13:36.066 "data_size": 65536 00:13:36.066 } 00:13:36.066 ] 00:13:36.066 }' 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:36.066 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.325 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:36.325 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:36.325 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:36.325 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:36.325 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:36.325 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.325 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.326 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.326 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.326 16:51:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.326 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:36.326 "name": "raid_bdev1", 00:13:36.326 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:36.326 "strip_size_kb": 64, 00:13:36.326 "state": "online", 00:13:36.326 "raid_level": "raid5f", 00:13:36.326 "superblock": false, 00:13:36.326 "num_base_bdevs": 3, 00:13:36.326 "num_base_bdevs_discovered": 2, 00:13:36.326 "num_base_bdevs_operational": 2, 00:13:36.326 "base_bdevs_list": [ 00:13:36.326 { 00:13:36.326 "name": null, 00:13:36.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.326 "is_configured": false, 00:13:36.326 "data_offset": 0, 00:13:36.326 "data_size": 65536 00:13:36.326 }, 00:13:36.326 { 00:13:36.326 "name": "BaseBdev2", 00:13:36.326 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:36.326 "is_configured": true, 00:13:36.326 "data_offset": 0, 00:13:36.326 "data_size": 65536 00:13:36.326 }, 00:13:36.326 { 00:13:36.326 "name": "BaseBdev3", 00:13:36.326 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:36.326 "is_configured": true, 00:13:36.326 "data_offset": 0, 00:13:36.326 "data_size": 65536 00:13:36.326 } 00:13:36.326 ] 00:13:36.326 }' 00:13:36.326 16:51:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:36.635 [2024-09-29 16:51:58.099613] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:36.635 [2024-09-29 16:51:58.103516] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.635 16:51:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:36.635 [2024-09-29 16:51:58.105776] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:37.575 "name": "raid_bdev1", 00:13:37.575 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:37.575 "strip_size_kb": 64, 00:13:37.575 "state": "online", 00:13:37.575 "raid_level": "raid5f", 00:13:37.575 "superblock": false, 00:13:37.575 "num_base_bdevs": 3, 00:13:37.575 "num_base_bdevs_discovered": 3, 00:13:37.575 "num_base_bdevs_operational": 3, 00:13:37.575 "process": { 00:13:37.575 "type": "rebuild", 00:13:37.575 "target": "spare", 00:13:37.575 "progress": { 00:13:37.575 "blocks": 20480, 00:13:37.575 "percent": 15 00:13:37.575 } 00:13:37.575 }, 00:13:37.575 "base_bdevs_list": [ 00:13:37.575 { 00:13:37.575 "name": "spare", 00:13:37.575 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:37.575 "is_configured": true, 00:13:37.575 "data_offset": 0, 00:13:37.575 "data_size": 65536 00:13:37.575 }, 00:13:37.575 { 00:13:37.575 "name": "BaseBdev2", 00:13:37.575 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:37.575 "is_configured": true, 00:13:37.575 "data_offset": 0, 00:13:37.575 "data_size": 65536 00:13:37.575 }, 00:13:37.575 { 00:13:37.575 "name": "BaseBdev3", 00:13:37.575 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:37.575 "is_configured": true, 00:13:37.575 "data_offset": 0, 00:13:37.575 "data_size": 65536 00:13:37.575 } 00:13:37.575 ] 00:13:37.575 }' 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:37.575 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=444 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:37.836 "name": "raid_bdev1", 00:13:37.836 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:37.836 "strip_size_kb": 64, 00:13:37.836 "state": "online", 00:13:37.836 "raid_level": "raid5f", 00:13:37.836 "superblock": false, 00:13:37.836 "num_base_bdevs": 3, 00:13:37.836 "num_base_bdevs_discovered": 3, 00:13:37.836 "num_base_bdevs_operational": 3, 00:13:37.836 "process": { 00:13:37.836 "type": "rebuild", 00:13:37.836 "target": "spare", 00:13:37.836 "progress": { 00:13:37.836 "blocks": 22528, 00:13:37.836 "percent": 17 00:13:37.836 } 00:13:37.836 }, 00:13:37.836 "base_bdevs_list": [ 00:13:37.836 { 00:13:37.836 "name": "spare", 00:13:37.836 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:37.836 "is_configured": true, 00:13:37.836 "data_offset": 0, 00:13:37.836 "data_size": 65536 00:13:37.836 }, 00:13:37.836 { 00:13:37.836 "name": "BaseBdev2", 00:13:37.836 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:37.836 "is_configured": true, 00:13:37.836 "data_offset": 0, 00:13:37.836 "data_size": 65536 00:13:37.836 }, 00:13:37.836 { 00:13:37.836 "name": "BaseBdev3", 00:13:37.836 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:37.836 "is_configured": true, 00:13:37.836 "data_offset": 0, 00:13:37.836 "data_size": 65536 00:13:37.836 } 00:13:37.836 ] 00:13:37.836 }' 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:37.836 16:51:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:38.774 16:52:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:39.033 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:39.033 "name": "raid_bdev1", 00:13:39.033 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:39.033 "strip_size_kb": 64, 00:13:39.033 "state": "online", 00:13:39.033 "raid_level": "raid5f", 00:13:39.033 "superblock": false, 00:13:39.033 "num_base_bdevs": 3, 00:13:39.033 "num_base_bdevs_discovered": 3, 00:13:39.033 "num_base_bdevs_operational": 3, 00:13:39.033 "process": { 00:13:39.033 "type": "rebuild", 00:13:39.033 "target": "spare", 00:13:39.033 "progress": { 00:13:39.033 "blocks": 45056, 00:13:39.033 "percent": 34 00:13:39.033 } 00:13:39.033 }, 00:13:39.033 "base_bdevs_list": [ 00:13:39.033 { 00:13:39.033 "name": "spare", 00:13:39.033 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:39.033 "is_configured": true, 00:13:39.033 "data_offset": 0, 00:13:39.033 "data_size": 65536 00:13:39.033 }, 00:13:39.033 { 00:13:39.033 "name": "BaseBdev2", 00:13:39.033 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:39.033 "is_configured": true, 00:13:39.033 "data_offset": 0, 00:13:39.033 "data_size": 65536 00:13:39.033 }, 00:13:39.033 { 00:13:39.033 "name": "BaseBdev3", 00:13:39.033 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:39.033 "is_configured": true, 00:13:39.033 "data_offset": 0, 00:13:39.033 "data_size": 65536 00:13:39.033 } 00:13:39.033 ] 00:13:39.033 }' 00:13:39.033 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:39.033 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:39.033 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:39.033 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:39.033 16:52:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:39.972 "name": "raid_bdev1", 00:13:39.972 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:39.972 "strip_size_kb": 64, 00:13:39.972 "state": "online", 00:13:39.972 "raid_level": "raid5f", 00:13:39.972 "superblock": false, 00:13:39.972 "num_base_bdevs": 3, 00:13:39.972 "num_base_bdevs_discovered": 3, 00:13:39.972 "num_base_bdevs_operational": 3, 00:13:39.972 "process": { 00:13:39.972 "type": "rebuild", 00:13:39.972 "target": "spare", 00:13:39.972 "progress": { 00:13:39.972 "blocks": 69632, 00:13:39.972 "percent": 53 00:13:39.972 } 00:13:39.972 }, 00:13:39.972 "base_bdevs_list": [ 00:13:39.972 { 00:13:39.972 "name": "spare", 00:13:39.972 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:39.972 "is_configured": true, 00:13:39.972 "data_offset": 0, 00:13:39.972 "data_size": 65536 00:13:39.972 }, 00:13:39.972 { 00:13:39.972 "name": "BaseBdev2", 00:13:39.972 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:39.972 "is_configured": true, 00:13:39.972 "data_offset": 0, 00:13:39.972 "data_size": 65536 00:13:39.972 }, 00:13:39.972 { 00:13:39.972 "name": "BaseBdev3", 00:13:39.972 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:39.972 "is_configured": true, 00:13:39.972 "data_offset": 0, 00:13:39.972 "data_size": 65536 00:13:39.972 } 00:13:39.972 ] 00:13:39.972 }' 00:13:39.972 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:40.231 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:40.231 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:40.231 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:40.231 16:52:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:41.169 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:41.169 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:41.170 "name": "raid_bdev1", 00:13:41.170 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:41.170 "strip_size_kb": 64, 00:13:41.170 "state": "online", 00:13:41.170 "raid_level": "raid5f", 00:13:41.170 "superblock": false, 00:13:41.170 "num_base_bdevs": 3, 00:13:41.170 "num_base_bdevs_discovered": 3, 00:13:41.170 "num_base_bdevs_operational": 3, 00:13:41.170 "process": { 00:13:41.170 "type": "rebuild", 00:13:41.170 "target": "spare", 00:13:41.170 "progress": { 00:13:41.170 "blocks": 92160, 00:13:41.170 "percent": 70 00:13:41.170 } 00:13:41.170 }, 00:13:41.170 "base_bdevs_list": [ 00:13:41.170 { 00:13:41.170 "name": "spare", 00:13:41.170 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:41.170 "is_configured": true, 00:13:41.170 "data_offset": 0, 00:13:41.170 "data_size": 65536 00:13:41.170 }, 00:13:41.170 { 00:13:41.170 "name": "BaseBdev2", 00:13:41.170 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:41.170 "is_configured": true, 00:13:41.170 "data_offset": 0, 00:13:41.170 "data_size": 65536 00:13:41.170 }, 00:13:41.170 { 00:13:41.170 "name": "BaseBdev3", 00:13:41.170 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:41.170 "is_configured": true, 00:13:41.170 "data_offset": 0, 00:13:41.170 "data_size": 65536 00:13:41.170 } 00:13:41.170 ] 00:13:41.170 }' 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:41.170 16:52:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:42.548 "name": "raid_bdev1", 00:13:42.548 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:42.548 "strip_size_kb": 64, 00:13:42.548 "state": "online", 00:13:42.548 "raid_level": "raid5f", 00:13:42.548 "superblock": false, 00:13:42.548 "num_base_bdevs": 3, 00:13:42.548 "num_base_bdevs_discovered": 3, 00:13:42.548 "num_base_bdevs_operational": 3, 00:13:42.548 "process": { 00:13:42.548 "type": "rebuild", 00:13:42.548 "target": "spare", 00:13:42.548 "progress": { 00:13:42.548 "blocks": 116736, 00:13:42.548 "percent": 89 00:13:42.548 } 00:13:42.548 }, 00:13:42.548 "base_bdevs_list": [ 00:13:42.548 { 00:13:42.548 "name": "spare", 00:13:42.548 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:42.548 "is_configured": true, 00:13:42.548 "data_offset": 0, 00:13:42.548 "data_size": 65536 00:13:42.548 }, 00:13:42.548 { 00:13:42.548 "name": "BaseBdev2", 00:13:42.548 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:42.548 "is_configured": true, 00:13:42.548 "data_offset": 0, 00:13:42.548 "data_size": 65536 00:13:42.548 }, 00:13:42.548 { 00:13:42.548 "name": "BaseBdev3", 00:13:42.548 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:42.548 "is_configured": true, 00:13:42.548 "data_offset": 0, 00:13:42.548 "data_size": 65536 00:13:42.548 } 00:13:42.548 ] 00:13:42.548 }' 00:13:42.548 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:42.549 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:42.549 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:42.549 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:42.549 16:52:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:43.117 [2024-09-29 16:52:04.539094] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:43.117 [2024-09-29 16:52:04.539170] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:43.117 [2024-09-29 16:52:04.539211] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.377 16:52:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.377 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.377 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:43.377 "name": "raid_bdev1", 00:13:43.377 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:43.377 "strip_size_kb": 64, 00:13:43.377 "state": "online", 00:13:43.377 "raid_level": "raid5f", 00:13:43.377 "superblock": false, 00:13:43.377 "num_base_bdevs": 3, 00:13:43.377 "num_base_bdevs_discovered": 3, 00:13:43.377 "num_base_bdevs_operational": 3, 00:13:43.377 "base_bdevs_list": [ 00:13:43.377 { 00:13:43.377 "name": "spare", 00:13:43.377 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:43.377 "is_configured": true, 00:13:43.377 "data_offset": 0, 00:13:43.377 "data_size": 65536 00:13:43.377 }, 00:13:43.377 { 00:13:43.377 "name": "BaseBdev2", 00:13:43.377 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:43.377 "is_configured": true, 00:13:43.377 "data_offset": 0, 00:13:43.377 "data_size": 65536 00:13:43.377 }, 00:13:43.377 { 00:13:43.377 "name": "BaseBdev3", 00:13:43.377 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:43.377 "is_configured": true, 00:13:43.377 "data_offset": 0, 00:13:43.377 "data_size": 65536 00:13:43.377 } 00:13:43.377 ] 00:13:43.377 }' 00:13:43.377 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:43.636 "name": "raid_bdev1", 00:13:43.636 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:43.636 "strip_size_kb": 64, 00:13:43.636 "state": "online", 00:13:43.636 "raid_level": "raid5f", 00:13:43.636 "superblock": false, 00:13:43.636 "num_base_bdevs": 3, 00:13:43.636 "num_base_bdevs_discovered": 3, 00:13:43.636 "num_base_bdevs_operational": 3, 00:13:43.636 "base_bdevs_list": [ 00:13:43.636 { 00:13:43.636 "name": "spare", 00:13:43.636 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:43.636 "is_configured": true, 00:13:43.636 "data_offset": 0, 00:13:43.636 "data_size": 65536 00:13:43.636 }, 00:13:43.636 { 00:13:43.636 "name": "BaseBdev2", 00:13:43.636 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:43.636 "is_configured": true, 00:13:43.636 "data_offset": 0, 00:13:43.636 "data_size": 65536 00:13:43.636 }, 00:13:43.636 { 00:13:43.636 "name": "BaseBdev3", 00:13:43.636 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:43.636 "is_configured": true, 00:13:43.636 "data_offset": 0, 00:13:43.636 "data_size": 65536 00:13:43.636 } 00:13:43.636 ] 00:13:43.636 }' 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.636 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:43.896 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.896 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.896 "name": "raid_bdev1", 00:13:43.896 "uuid": "75983574-df31-409b-aa5f-9879464c31b2", 00:13:43.896 "strip_size_kb": 64, 00:13:43.896 "state": "online", 00:13:43.896 "raid_level": "raid5f", 00:13:43.896 "superblock": false, 00:13:43.896 "num_base_bdevs": 3, 00:13:43.896 "num_base_bdevs_discovered": 3, 00:13:43.896 "num_base_bdevs_operational": 3, 00:13:43.896 "base_bdevs_list": [ 00:13:43.896 { 00:13:43.896 "name": "spare", 00:13:43.896 "uuid": "2a0ed790-0032-5fee-8140-173e053d5235", 00:13:43.896 "is_configured": true, 00:13:43.896 "data_offset": 0, 00:13:43.896 "data_size": 65536 00:13:43.896 }, 00:13:43.896 { 00:13:43.896 "name": "BaseBdev2", 00:13:43.896 "uuid": "904e8182-9b05-5ffe-8591-7b98e2dbf6dc", 00:13:43.896 "is_configured": true, 00:13:43.896 "data_offset": 0, 00:13:43.896 "data_size": 65536 00:13:43.896 }, 00:13:43.896 { 00:13:43.896 "name": "BaseBdev3", 00:13:43.896 "uuid": "936edd8a-5ae4-50b5-8b92-bb891a5375ef", 00:13:43.896 "is_configured": true, 00:13:43.896 "data_offset": 0, 00:13:43.896 "data_size": 65536 00:13:43.896 } 00:13:43.896 ] 00:13:43.896 }' 00:13:43.896 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.896 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.156 [2024-09-29 16:52:05.722196] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:44.156 [2024-09-29 16:52:05.722277] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:44.156 [2024-09-29 16:52:05.722457] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:44.156 [2024-09-29 16:52:05.722584] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:44.156 [2024-09-29 16:52:05.722631] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:44.156 16:52:05 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:44.416 /dev/nbd0 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:44.416 1+0 records in 00:13:44.416 1+0 records out 00:13:44.416 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000469889 s, 8.7 MB/s 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:44.416 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:44.676 /dev/nbd1 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:44.676 1+0 records in 00:13:44.676 1+0 records out 00:13:44.676 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000560807 s, 7.3 MB/s 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:44.676 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:44.936 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 91735 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 91735 ']' 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 91735 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91735 00:13:45.196 killing process with pid 91735 00:13:45.196 Received shutdown signal, test time was about 60.000000 seconds 00:13:45.196 00:13:45.196 Latency(us) 00:13:45.196 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:45.196 =================================================================================================================== 00:13:45.196 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91735' 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 91735 00:13:45.196 [2024-09-29 16:52:06.866734] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:45.196 16:52:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 91735 00:13:45.456 [2024-09-29 16:52:06.907553] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:13:45.716 00:13:45.716 real 0m13.702s 00:13:45.716 user 0m17.171s 00:13:45.716 sys 0m2.054s 00:13:45.716 ************************************ 00:13:45.716 END TEST raid5f_rebuild_test 00:13:45.716 ************************************ 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:13:45.716 16:52:07 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:13:45.716 16:52:07 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:13:45.716 16:52:07 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:45.716 16:52:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:45.716 ************************************ 00:13:45.716 START TEST raid5f_rebuild_test_sb 00:13:45.716 ************************************ 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 true false true 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92159 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92159 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 92159 ']' 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:45.716 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:45.716 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:45.717 16:52:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:45.717 [2024-09-29 16:52:07.324557] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:13:45.717 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:45.717 Zero copy mechanism will not be used. 00:13:45.717 [2024-09-29 16:52:07.324828] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid92159 ] 00:13:45.976 [2024-09-29 16:52:07.471386] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:45.976 [2024-09-29 16:52:07.520502] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:45.976 [2024-09-29 16:52:07.564267] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:45.976 [2024-09-29 16:52:07.564384] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.546 BaseBdev1_malloc 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.546 [2024-09-29 16:52:08.171121] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:46.546 [2024-09-29 16:52:08.171243] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:46.546 [2024-09-29 16:52:08.171289] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:46.546 [2024-09-29 16:52:08.171335] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:46.546 [2024-09-29 16:52:08.173477] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:46.546 [2024-09-29 16:52:08.173550] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:46.546 BaseBdev1 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.546 BaseBdev2_malloc 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.546 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.546 [2024-09-29 16:52:08.213804] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:46.546 [2024-09-29 16:52:08.214066] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:46.546 [2024-09-29 16:52:08.214169] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:46.546 [2024-09-29 16:52:08.214271] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:46.807 [2024-09-29 16:52:08.218937] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:46.807 [2024-09-29 16:52:08.219090] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:46.807 BaseBdev2 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.807 BaseBdev3_malloc 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.807 [2024-09-29 16:52:08.245356] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:46.807 [2024-09-29 16:52:08.245478] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:46.807 [2024-09-29 16:52:08.245525] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:46.807 [2024-09-29 16:52:08.245560] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:46.807 [2024-09-29 16:52:08.247736] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:46.807 [2024-09-29 16:52:08.247817] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:46.807 BaseBdev3 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.807 spare_malloc 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.807 spare_delay 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.807 [2024-09-29 16:52:08.286247] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:46.807 [2024-09-29 16:52:08.286337] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:46.807 [2024-09-29 16:52:08.286366] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:13:46.807 [2024-09-29 16:52:08.286374] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:46.807 [2024-09-29 16:52:08.288468] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:46.807 [2024-09-29 16:52:08.288573] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:46.807 spare 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.807 [2024-09-29 16:52:08.298291] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:46.807 [2024-09-29 16:52:08.300137] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:46.807 [2024-09-29 16:52:08.300263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:46.807 [2024-09-29 16:52:08.300478] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:46.807 [2024-09-29 16:52:08.300534] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:46.807 [2024-09-29 16:52:08.300812] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:13:46.807 [2024-09-29 16:52:08.301260] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:46.807 [2024-09-29 16:52:08.301308] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:46.807 [2024-09-29 16:52:08.301490] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:46.807 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:46.808 "name": "raid_bdev1", 00:13:46.808 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:46.808 "strip_size_kb": 64, 00:13:46.808 "state": "online", 00:13:46.808 "raid_level": "raid5f", 00:13:46.808 "superblock": true, 00:13:46.808 "num_base_bdevs": 3, 00:13:46.808 "num_base_bdevs_discovered": 3, 00:13:46.808 "num_base_bdevs_operational": 3, 00:13:46.808 "base_bdevs_list": [ 00:13:46.808 { 00:13:46.808 "name": "BaseBdev1", 00:13:46.808 "uuid": "397cecd9-3826-50ca-9d14-73641a97d357", 00:13:46.808 "is_configured": true, 00:13:46.808 "data_offset": 2048, 00:13:46.808 "data_size": 63488 00:13:46.808 }, 00:13:46.808 { 00:13:46.808 "name": "BaseBdev2", 00:13:46.808 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:46.808 "is_configured": true, 00:13:46.808 "data_offset": 2048, 00:13:46.808 "data_size": 63488 00:13:46.808 }, 00:13:46.808 { 00:13:46.808 "name": "BaseBdev3", 00:13:46.808 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:46.808 "is_configured": true, 00:13:46.808 "data_offset": 2048, 00:13:46.808 "data_size": 63488 00:13:46.808 } 00:13:46.808 ] 00:13:46.808 }' 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:46.808 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.068 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:47.068 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.068 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.068 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:47.068 [2024-09-29 16:52:08.738347] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:47.328 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:47.329 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:47.329 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:47.329 16:52:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:13:47.588 [2024-09-29 16:52:09.021706] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:13:47.588 /dev/nbd0 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:47.588 1+0 records in 00:13:47.588 1+0 records out 00:13:47.588 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000492229 s, 8.3 MB/s 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:13:47.588 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:13:47.849 496+0 records in 00:13:47.849 496+0 records out 00:13:47.849 65011712 bytes (65 MB, 62 MiB) copied, 0.309163 s, 210 MB/s 00:13:47.849 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:47.849 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:47.849 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:47.849 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:47.849 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:47.849 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:47.849 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:48.109 [2024-09-29 16:52:09.626387] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.109 [2024-09-29 16:52:09.634453] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.109 "name": "raid_bdev1", 00:13:48.109 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:48.109 "strip_size_kb": 64, 00:13:48.109 "state": "online", 00:13:48.109 "raid_level": "raid5f", 00:13:48.109 "superblock": true, 00:13:48.109 "num_base_bdevs": 3, 00:13:48.109 "num_base_bdevs_discovered": 2, 00:13:48.109 "num_base_bdevs_operational": 2, 00:13:48.109 "base_bdevs_list": [ 00:13:48.109 { 00:13:48.109 "name": null, 00:13:48.109 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.109 "is_configured": false, 00:13:48.109 "data_offset": 0, 00:13:48.109 "data_size": 63488 00:13:48.109 }, 00:13:48.109 { 00:13:48.109 "name": "BaseBdev2", 00:13:48.109 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:48.109 "is_configured": true, 00:13:48.109 "data_offset": 2048, 00:13:48.109 "data_size": 63488 00:13:48.109 }, 00:13:48.109 { 00:13:48.109 "name": "BaseBdev3", 00:13:48.109 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:48.109 "is_configured": true, 00:13:48.109 "data_offset": 2048, 00:13:48.109 "data_size": 63488 00:13:48.109 } 00:13:48.109 ] 00:13:48.109 }' 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.109 16:52:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.679 16:52:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:48.679 16:52:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.679 16:52:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:48.679 [2024-09-29 16:52:10.057786] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:48.679 [2024-09-29 16:52:10.061684] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000255d0 00:13:48.679 16:52:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.679 16:52:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:48.679 [2024-09-29 16:52:10.063910] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:49.620 "name": "raid_bdev1", 00:13:49.620 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:49.620 "strip_size_kb": 64, 00:13:49.620 "state": "online", 00:13:49.620 "raid_level": "raid5f", 00:13:49.620 "superblock": true, 00:13:49.620 "num_base_bdevs": 3, 00:13:49.620 "num_base_bdevs_discovered": 3, 00:13:49.620 "num_base_bdevs_operational": 3, 00:13:49.620 "process": { 00:13:49.620 "type": "rebuild", 00:13:49.620 "target": "spare", 00:13:49.620 "progress": { 00:13:49.620 "blocks": 20480, 00:13:49.620 "percent": 16 00:13:49.620 } 00:13:49.620 }, 00:13:49.620 "base_bdevs_list": [ 00:13:49.620 { 00:13:49.620 "name": "spare", 00:13:49.620 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:49.620 "is_configured": true, 00:13:49.620 "data_offset": 2048, 00:13:49.620 "data_size": 63488 00:13:49.620 }, 00:13:49.620 { 00:13:49.620 "name": "BaseBdev2", 00:13:49.620 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:49.620 "is_configured": true, 00:13:49.620 "data_offset": 2048, 00:13:49.620 "data_size": 63488 00:13:49.620 }, 00:13:49.620 { 00:13:49.620 "name": "BaseBdev3", 00:13:49.620 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:49.620 "is_configured": true, 00:13:49.620 "data_offset": 2048, 00:13:49.620 "data_size": 63488 00:13:49.620 } 00:13:49.620 ] 00:13:49.620 }' 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.620 [2024-09-29 16:52:11.192821] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:49.620 [2024-09-29 16:52:11.270902] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:49.620 [2024-09-29 16:52:11.271002] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:49.620 [2024-09-29 16:52:11.271036] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:49.620 [2024-09-29 16:52:11.271089] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:49.620 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:49.880 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:49.880 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:49.880 "name": "raid_bdev1", 00:13:49.880 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:49.880 "strip_size_kb": 64, 00:13:49.880 "state": "online", 00:13:49.880 "raid_level": "raid5f", 00:13:49.880 "superblock": true, 00:13:49.880 "num_base_bdevs": 3, 00:13:49.880 "num_base_bdevs_discovered": 2, 00:13:49.880 "num_base_bdevs_operational": 2, 00:13:49.880 "base_bdevs_list": [ 00:13:49.880 { 00:13:49.880 "name": null, 00:13:49.880 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:49.880 "is_configured": false, 00:13:49.880 "data_offset": 0, 00:13:49.880 "data_size": 63488 00:13:49.880 }, 00:13:49.880 { 00:13:49.880 "name": "BaseBdev2", 00:13:49.880 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:49.880 "is_configured": true, 00:13:49.880 "data_offset": 2048, 00:13:49.880 "data_size": 63488 00:13:49.880 }, 00:13:49.880 { 00:13:49.880 "name": "BaseBdev3", 00:13:49.880 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:49.880 "is_configured": true, 00:13:49.880 "data_offset": 2048, 00:13:49.880 "data_size": 63488 00:13:49.880 } 00:13:49.880 ] 00:13:49.880 }' 00:13:49.880 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:49.880 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.139 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:50.399 "name": "raid_bdev1", 00:13:50.399 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:50.399 "strip_size_kb": 64, 00:13:50.399 "state": "online", 00:13:50.399 "raid_level": "raid5f", 00:13:50.399 "superblock": true, 00:13:50.399 "num_base_bdevs": 3, 00:13:50.399 "num_base_bdevs_discovered": 2, 00:13:50.399 "num_base_bdevs_operational": 2, 00:13:50.399 "base_bdevs_list": [ 00:13:50.399 { 00:13:50.399 "name": null, 00:13:50.399 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.399 "is_configured": false, 00:13:50.399 "data_offset": 0, 00:13:50.399 "data_size": 63488 00:13:50.399 }, 00:13:50.399 { 00:13:50.399 "name": "BaseBdev2", 00:13:50.399 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:50.399 "is_configured": true, 00:13:50.399 "data_offset": 2048, 00:13:50.399 "data_size": 63488 00:13:50.399 }, 00:13:50.399 { 00:13:50.399 "name": "BaseBdev3", 00:13:50.399 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:50.399 "is_configured": true, 00:13:50.399 "data_offset": 2048, 00:13:50.399 "data_size": 63488 00:13:50.399 } 00:13:50.399 ] 00:13:50.399 }' 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:50.399 [2024-09-29 16:52:11.891462] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:50.399 [2024-09-29 16:52:11.894995] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000256a0 00:13:50.399 [2024-09-29 16:52:11.897127] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.399 16:52:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.338 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:51.338 "name": "raid_bdev1", 00:13:51.338 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:51.338 "strip_size_kb": 64, 00:13:51.338 "state": "online", 00:13:51.338 "raid_level": "raid5f", 00:13:51.338 "superblock": true, 00:13:51.338 "num_base_bdevs": 3, 00:13:51.338 "num_base_bdevs_discovered": 3, 00:13:51.338 "num_base_bdevs_operational": 3, 00:13:51.338 "process": { 00:13:51.338 "type": "rebuild", 00:13:51.338 "target": "spare", 00:13:51.338 "progress": { 00:13:51.338 "blocks": 20480, 00:13:51.338 "percent": 16 00:13:51.338 } 00:13:51.338 }, 00:13:51.339 "base_bdevs_list": [ 00:13:51.339 { 00:13:51.339 "name": "spare", 00:13:51.339 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:51.339 "is_configured": true, 00:13:51.339 "data_offset": 2048, 00:13:51.339 "data_size": 63488 00:13:51.339 }, 00:13:51.339 { 00:13:51.339 "name": "BaseBdev2", 00:13:51.339 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:51.339 "is_configured": true, 00:13:51.339 "data_offset": 2048, 00:13:51.339 "data_size": 63488 00:13:51.339 }, 00:13:51.339 { 00:13:51.339 "name": "BaseBdev3", 00:13:51.339 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:51.339 "is_configured": true, 00:13:51.339 "data_offset": 2048, 00:13:51.339 "data_size": 63488 00:13:51.339 } 00:13:51.339 ] 00:13:51.339 }' 00:13:51.339 16:52:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:51.339 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:51.339 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:51.599 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=458 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:51.599 "name": "raid_bdev1", 00:13:51.599 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:51.599 "strip_size_kb": 64, 00:13:51.599 "state": "online", 00:13:51.599 "raid_level": "raid5f", 00:13:51.599 "superblock": true, 00:13:51.599 "num_base_bdevs": 3, 00:13:51.599 "num_base_bdevs_discovered": 3, 00:13:51.599 "num_base_bdevs_operational": 3, 00:13:51.599 "process": { 00:13:51.599 "type": "rebuild", 00:13:51.599 "target": "spare", 00:13:51.599 "progress": { 00:13:51.599 "blocks": 22528, 00:13:51.599 "percent": 17 00:13:51.599 } 00:13:51.599 }, 00:13:51.599 "base_bdevs_list": [ 00:13:51.599 { 00:13:51.599 "name": "spare", 00:13:51.599 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:51.599 "is_configured": true, 00:13:51.599 "data_offset": 2048, 00:13:51.599 "data_size": 63488 00:13:51.599 }, 00:13:51.599 { 00:13:51.599 "name": "BaseBdev2", 00:13:51.599 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:51.599 "is_configured": true, 00:13:51.599 "data_offset": 2048, 00:13:51.599 "data_size": 63488 00:13:51.599 }, 00:13:51.599 { 00:13:51.599 "name": "BaseBdev3", 00:13:51.599 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:51.599 "is_configured": true, 00:13:51.599 "data_offset": 2048, 00:13:51.599 "data_size": 63488 00:13:51.599 } 00:13:51.599 ] 00:13:51.599 }' 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:51.599 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:51.600 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:51.600 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:51.600 16:52:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:52.540 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.541 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:52.541 "name": "raid_bdev1", 00:13:52.541 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:52.541 "strip_size_kb": 64, 00:13:52.541 "state": "online", 00:13:52.541 "raid_level": "raid5f", 00:13:52.541 "superblock": true, 00:13:52.541 "num_base_bdevs": 3, 00:13:52.541 "num_base_bdevs_discovered": 3, 00:13:52.541 "num_base_bdevs_operational": 3, 00:13:52.541 "process": { 00:13:52.541 "type": "rebuild", 00:13:52.541 "target": "spare", 00:13:52.541 "progress": { 00:13:52.541 "blocks": 45056, 00:13:52.541 "percent": 35 00:13:52.541 } 00:13:52.541 }, 00:13:52.541 "base_bdevs_list": [ 00:13:52.541 { 00:13:52.541 "name": "spare", 00:13:52.541 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:52.541 "is_configured": true, 00:13:52.541 "data_offset": 2048, 00:13:52.541 "data_size": 63488 00:13:52.541 }, 00:13:52.541 { 00:13:52.541 "name": "BaseBdev2", 00:13:52.541 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:52.541 "is_configured": true, 00:13:52.541 "data_offset": 2048, 00:13:52.541 "data_size": 63488 00:13:52.541 }, 00:13:52.541 { 00:13:52.541 "name": "BaseBdev3", 00:13:52.541 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:52.541 "is_configured": true, 00:13:52.541 "data_offset": 2048, 00:13:52.541 "data_size": 63488 00:13:52.541 } 00:13:52.541 ] 00:13:52.541 }' 00:13:52.800 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:52.800 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:52.801 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:52.801 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:52.801 16:52:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:53.737 "name": "raid_bdev1", 00:13:53.737 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:53.737 "strip_size_kb": 64, 00:13:53.737 "state": "online", 00:13:53.737 "raid_level": "raid5f", 00:13:53.737 "superblock": true, 00:13:53.737 "num_base_bdevs": 3, 00:13:53.737 "num_base_bdevs_discovered": 3, 00:13:53.737 "num_base_bdevs_operational": 3, 00:13:53.737 "process": { 00:13:53.737 "type": "rebuild", 00:13:53.737 "target": "spare", 00:13:53.737 "progress": { 00:13:53.737 "blocks": 69632, 00:13:53.737 "percent": 54 00:13:53.737 } 00:13:53.737 }, 00:13:53.737 "base_bdevs_list": [ 00:13:53.737 { 00:13:53.737 "name": "spare", 00:13:53.737 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:53.737 "is_configured": true, 00:13:53.737 "data_offset": 2048, 00:13:53.737 "data_size": 63488 00:13:53.737 }, 00:13:53.737 { 00:13:53.737 "name": "BaseBdev2", 00:13:53.737 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:53.737 "is_configured": true, 00:13:53.737 "data_offset": 2048, 00:13:53.737 "data_size": 63488 00:13:53.737 }, 00:13:53.737 { 00:13:53.737 "name": "BaseBdev3", 00:13:53.737 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:53.737 "is_configured": true, 00:13:53.737 "data_offset": 2048, 00:13:53.737 "data_size": 63488 00:13:53.737 } 00:13:53.737 ] 00:13:53.737 }' 00:13:53.737 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:53.996 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:53.996 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:53.996 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:53.996 16:52:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:54.931 "name": "raid_bdev1", 00:13:54.931 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:54.931 "strip_size_kb": 64, 00:13:54.931 "state": "online", 00:13:54.931 "raid_level": "raid5f", 00:13:54.931 "superblock": true, 00:13:54.931 "num_base_bdevs": 3, 00:13:54.931 "num_base_bdevs_discovered": 3, 00:13:54.931 "num_base_bdevs_operational": 3, 00:13:54.931 "process": { 00:13:54.931 "type": "rebuild", 00:13:54.931 "target": "spare", 00:13:54.931 "progress": { 00:13:54.931 "blocks": 92160, 00:13:54.931 "percent": 72 00:13:54.931 } 00:13:54.931 }, 00:13:54.931 "base_bdevs_list": [ 00:13:54.931 { 00:13:54.931 "name": "spare", 00:13:54.931 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:54.931 "is_configured": true, 00:13:54.931 "data_offset": 2048, 00:13:54.931 "data_size": 63488 00:13:54.931 }, 00:13:54.931 { 00:13:54.931 "name": "BaseBdev2", 00:13:54.931 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:54.931 "is_configured": true, 00:13:54.931 "data_offset": 2048, 00:13:54.931 "data_size": 63488 00:13:54.931 }, 00:13:54.931 { 00:13:54.931 "name": "BaseBdev3", 00:13:54.931 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:54.931 "is_configured": true, 00:13:54.931 "data_offset": 2048, 00:13:54.931 "data_size": 63488 00:13:54.931 } 00:13:54.931 ] 00:13:54.931 }' 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:54.931 16:52:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:56.305 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.306 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:56.306 "name": "raid_bdev1", 00:13:56.306 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:56.306 "strip_size_kb": 64, 00:13:56.306 "state": "online", 00:13:56.306 "raid_level": "raid5f", 00:13:56.306 "superblock": true, 00:13:56.306 "num_base_bdevs": 3, 00:13:56.306 "num_base_bdevs_discovered": 3, 00:13:56.306 "num_base_bdevs_operational": 3, 00:13:56.306 "process": { 00:13:56.306 "type": "rebuild", 00:13:56.306 "target": "spare", 00:13:56.306 "progress": { 00:13:56.306 "blocks": 114688, 00:13:56.306 "percent": 90 00:13:56.306 } 00:13:56.306 }, 00:13:56.306 "base_bdevs_list": [ 00:13:56.306 { 00:13:56.306 "name": "spare", 00:13:56.306 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:56.306 "is_configured": true, 00:13:56.306 "data_offset": 2048, 00:13:56.306 "data_size": 63488 00:13:56.306 }, 00:13:56.306 { 00:13:56.306 "name": "BaseBdev2", 00:13:56.306 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:56.306 "is_configured": true, 00:13:56.306 "data_offset": 2048, 00:13:56.306 "data_size": 63488 00:13:56.306 }, 00:13:56.306 { 00:13:56.306 "name": "BaseBdev3", 00:13:56.306 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:56.306 "is_configured": true, 00:13:56.306 "data_offset": 2048, 00:13:56.306 "data_size": 63488 00:13:56.306 } 00:13:56.306 ] 00:13:56.306 }' 00:13:56.306 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:56.306 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:56.306 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:56.306 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:56.306 16:52:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:56.565 [2024-09-29 16:52:18.129958] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:56.565 [2024-09-29 16:52:18.130075] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:56.565 [2024-09-29 16:52:18.130191] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:57.134 "name": "raid_bdev1", 00:13:57.134 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:57.134 "strip_size_kb": 64, 00:13:57.134 "state": "online", 00:13:57.134 "raid_level": "raid5f", 00:13:57.134 "superblock": true, 00:13:57.134 "num_base_bdevs": 3, 00:13:57.134 "num_base_bdevs_discovered": 3, 00:13:57.134 "num_base_bdevs_operational": 3, 00:13:57.134 "base_bdevs_list": [ 00:13:57.134 { 00:13:57.134 "name": "spare", 00:13:57.134 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:57.134 "is_configured": true, 00:13:57.134 "data_offset": 2048, 00:13:57.134 "data_size": 63488 00:13:57.134 }, 00:13:57.134 { 00:13:57.134 "name": "BaseBdev2", 00:13:57.134 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:57.134 "is_configured": true, 00:13:57.134 "data_offset": 2048, 00:13:57.134 "data_size": 63488 00:13:57.134 }, 00:13:57.134 { 00:13:57.134 "name": "BaseBdev3", 00:13:57.134 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:57.134 "is_configured": true, 00:13:57.134 "data_offset": 2048, 00:13:57.134 "data_size": 63488 00:13:57.134 } 00:13:57.134 ] 00:13:57.134 }' 00:13:57.134 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.394 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:57.394 "name": "raid_bdev1", 00:13:57.394 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:57.394 "strip_size_kb": 64, 00:13:57.394 "state": "online", 00:13:57.394 "raid_level": "raid5f", 00:13:57.394 "superblock": true, 00:13:57.394 "num_base_bdevs": 3, 00:13:57.394 "num_base_bdevs_discovered": 3, 00:13:57.394 "num_base_bdevs_operational": 3, 00:13:57.394 "base_bdevs_list": [ 00:13:57.394 { 00:13:57.394 "name": "spare", 00:13:57.394 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:57.394 "is_configured": true, 00:13:57.394 "data_offset": 2048, 00:13:57.394 "data_size": 63488 00:13:57.394 }, 00:13:57.394 { 00:13:57.394 "name": "BaseBdev2", 00:13:57.394 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:57.394 "is_configured": true, 00:13:57.394 "data_offset": 2048, 00:13:57.394 "data_size": 63488 00:13:57.394 }, 00:13:57.394 { 00:13:57.394 "name": "BaseBdev3", 00:13:57.395 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:57.395 "is_configured": true, 00:13:57.395 "data_offset": 2048, 00:13:57.395 "data_size": 63488 00:13:57.395 } 00:13:57.395 ] 00:13:57.395 }' 00:13:57.395 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:57.395 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:57.395 16:52:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.395 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.654 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.654 "name": "raid_bdev1", 00:13:57.654 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:57.654 "strip_size_kb": 64, 00:13:57.654 "state": "online", 00:13:57.654 "raid_level": "raid5f", 00:13:57.654 "superblock": true, 00:13:57.654 "num_base_bdevs": 3, 00:13:57.654 "num_base_bdevs_discovered": 3, 00:13:57.654 "num_base_bdevs_operational": 3, 00:13:57.654 "base_bdevs_list": [ 00:13:57.654 { 00:13:57.654 "name": "spare", 00:13:57.654 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:57.654 "is_configured": true, 00:13:57.654 "data_offset": 2048, 00:13:57.654 "data_size": 63488 00:13:57.654 }, 00:13:57.654 { 00:13:57.654 "name": "BaseBdev2", 00:13:57.654 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:57.654 "is_configured": true, 00:13:57.654 "data_offset": 2048, 00:13:57.654 "data_size": 63488 00:13:57.654 }, 00:13:57.654 { 00:13:57.654 "name": "BaseBdev3", 00:13:57.654 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:57.654 "is_configured": true, 00:13:57.654 "data_offset": 2048, 00:13:57.655 "data_size": 63488 00:13:57.655 } 00:13:57.655 ] 00:13:57.655 }' 00:13:57.655 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.655 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.914 [2024-09-29 16:52:19.468832] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:57.914 [2024-09-29 16:52:19.468867] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:57.914 [2024-09-29 16:52:19.468957] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:57.914 [2024-09-29 16:52:19.469033] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:57.914 [2024-09-29 16:52:19.469042] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:57.914 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:57.915 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:58.174 /dev/nbd0 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:58.174 1+0 records in 00:13:58.174 1+0 records out 00:13:58.174 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000562526 s, 7.3 MB/s 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:58.174 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:13:58.175 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:58.175 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:58.175 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:13:58.175 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:58.175 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:58.175 16:52:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:58.435 /dev/nbd1 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:58.435 1+0 records in 00:13:58.435 1+0 records out 00:13:58.435 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000446762 s, 9.2 MB/s 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:58.435 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:58.695 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.955 [2024-09-29 16:52:20.575432] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:58.955 [2024-09-29 16:52:20.575537] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:58.955 [2024-09-29 16:52:20.575601] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:13:58.955 [2024-09-29 16:52:20.575635] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:58.955 [2024-09-29 16:52:20.577873] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:58.955 [2024-09-29 16:52:20.577971] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:58.955 [2024-09-29 16:52:20.578068] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:58.955 [2024-09-29 16:52:20.578143] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:58.955 [2024-09-29 16:52:20.578316] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:58.955 [2024-09-29 16:52:20.578470] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:58.955 spare 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:58.955 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.215 [2024-09-29 16:52:20.678405] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:59.215 [2024-09-29 16:52:20.678461] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:13:59.215 [2024-09-29 16:52:20.678773] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043d50 00:13:59.215 [2024-09-29 16:52:20.679236] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:59.215 [2024-09-29 16:52:20.679288] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:59.215 [2024-09-29 16:52:20.679486] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:59.215 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.215 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:13:59.215 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:59.215 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.216 "name": "raid_bdev1", 00:13:59.216 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:59.216 "strip_size_kb": 64, 00:13:59.216 "state": "online", 00:13:59.216 "raid_level": "raid5f", 00:13:59.216 "superblock": true, 00:13:59.216 "num_base_bdevs": 3, 00:13:59.216 "num_base_bdevs_discovered": 3, 00:13:59.216 "num_base_bdevs_operational": 3, 00:13:59.216 "base_bdevs_list": [ 00:13:59.216 { 00:13:59.216 "name": "spare", 00:13:59.216 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:59.216 "is_configured": true, 00:13:59.216 "data_offset": 2048, 00:13:59.216 "data_size": 63488 00:13:59.216 }, 00:13:59.216 { 00:13:59.216 "name": "BaseBdev2", 00:13:59.216 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:59.216 "is_configured": true, 00:13:59.216 "data_offset": 2048, 00:13:59.216 "data_size": 63488 00:13:59.216 }, 00:13:59.216 { 00:13:59.216 "name": "BaseBdev3", 00:13:59.216 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:59.216 "is_configured": true, 00:13:59.216 "data_offset": 2048, 00:13:59.216 "data_size": 63488 00:13:59.216 } 00:13:59.216 ] 00:13:59.216 }' 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.216 16:52:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.786 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:59.787 "name": "raid_bdev1", 00:13:59.787 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:59.787 "strip_size_kb": 64, 00:13:59.787 "state": "online", 00:13:59.787 "raid_level": "raid5f", 00:13:59.787 "superblock": true, 00:13:59.787 "num_base_bdevs": 3, 00:13:59.787 "num_base_bdevs_discovered": 3, 00:13:59.787 "num_base_bdevs_operational": 3, 00:13:59.787 "base_bdevs_list": [ 00:13:59.787 { 00:13:59.787 "name": "spare", 00:13:59.787 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:13:59.787 "is_configured": true, 00:13:59.787 "data_offset": 2048, 00:13:59.787 "data_size": 63488 00:13:59.787 }, 00:13:59.787 { 00:13:59.787 "name": "BaseBdev2", 00:13:59.787 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:59.787 "is_configured": true, 00:13:59.787 "data_offset": 2048, 00:13:59.787 "data_size": 63488 00:13:59.787 }, 00:13:59.787 { 00:13:59.787 "name": "BaseBdev3", 00:13:59.787 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:59.787 "is_configured": true, 00:13:59.787 "data_offset": 2048, 00:13:59.787 "data_size": 63488 00:13:59.787 } 00:13:59.787 ] 00:13:59.787 }' 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.787 [2024-09-29 16:52:21.366353] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.787 "name": "raid_bdev1", 00:13:59.787 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:13:59.787 "strip_size_kb": 64, 00:13:59.787 "state": "online", 00:13:59.787 "raid_level": "raid5f", 00:13:59.787 "superblock": true, 00:13:59.787 "num_base_bdevs": 3, 00:13:59.787 "num_base_bdevs_discovered": 2, 00:13:59.787 "num_base_bdevs_operational": 2, 00:13:59.787 "base_bdevs_list": [ 00:13:59.787 { 00:13:59.787 "name": null, 00:13:59.787 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.787 "is_configured": false, 00:13:59.787 "data_offset": 0, 00:13:59.787 "data_size": 63488 00:13:59.787 }, 00:13:59.787 { 00:13:59.787 "name": "BaseBdev2", 00:13:59.787 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:13:59.787 "is_configured": true, 00:13:59.787 "data_offset": 2048, 00:13:59.787 "data_size": 63488 00:13:59.787 }, 00:13:59.787 { 00:13:59.787 "name": "BaseBdev3", 00:13:59.787 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:13:59.787 "is_configured": true, 00:13:59.787 "data_offset": 2048, 00:13:59.787 "data_size": 63488 00:13:59.787 } 00:13:59.787 ] 00:13:59.787 }' 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.787 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.357 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:00.357 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.357 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.357 [2024-09-29 16:52:21.825626] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:00.357 [2024-09-29 16:52:21.825867] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:00.357 [2024-09-29 16:52:21.825887] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:00.357 [2024-09-29 16:52:21.825924] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:00.357 [2024-09-29 16:52:21.829634] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043e20 00:14:00.357 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.357 16:52:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:00.357 [2024-09-29 16:52:21.831920] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.296 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:01.296 "name": "raid_bdev1", 00:14:01.296 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:01.296 "strip_size_kb": 64, 00:14:01.296 "state": "online", 00:14:01.296 "raid_level": "raid5f", 00:14:01.296 "superblock": true, 00:14:01.296 "num_base_bdevs": 3, 00:14:01.296 "num_base_bdevs_discovered": 3, 00:14:01.296 "num_base_bdevs_operational": 3, 00:14:01.296 "process": { 00:14:01.296 "type": "rebuild", 00:14:01.296 "target": "spare", 00:14:01.296 "progress": { 00:14:01.296 "blocks": 20480, 00:14:01.296 "percent": 16 00:14:01.296 } 00:14:01.296 }, 00:14:01.296 "base_bdevs_list": [ 00:14:01.296 { 00:14:01.296 "name": "spare", 00:14:01.296 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:14:01.296 "is_configured": true, 00:14:01.296 "data_offset": 2048, 00:14:01.296 "data_size": 63488 00:14:01.296 }, 00:14:01.296 { 00:14:01.296 "name": "BaseBdev2", 00:14:01.296 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:01.296 "is_configured": true, 00:14:01.296 "data_offset": 2048, 00:14:01.296 "data_size": 63488 00:14:01.296 }, 00:14:01.296 { 00:14:01.296 "name": "BaseBdev3", 00:14:01.296 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:01.296 "is_configured": true, 00:14:01.296 "data_offset": 2048, 00:14:01.296 "data_size": 63488 00:14:01.296 } 00:14:01.297 ] 00:14:01.297 }' 00:14:01.297 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:01.297 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:01.297 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:01.557 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:01.557 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:01.557 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.557 16:52:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.557 [2024-09-29 16:52:22.992231] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:01.557 [2024-09-29 16:52:23.038618] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:01.557 [2024-09-29 16:52:23.038727] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:01.557 [2024-09-29 16:52:23.038799] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:01.557 [2024-09-29 16:52:23.038847] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.557 "name": "raid_bdev1", 00:14:01.557 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:01.557 "strip_size_kb": 64, 00:14:01.557 "state": "online", 00:14:01.557 "raid_level": "raid5f", 00:14:01.557 "superblock": true, 00:14:01.557 "num_base_bdevs": 3, 00:14:01.557 "num_base_bdevs_discovered": 2, 00:14:01.557 "num_base_bdevs_operational": 2, 00:14:01.557 "base_bdevs_list": [ 00:14:01.557 { 00:14:01.557 "name": null, 00:14:01.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.557 "is_configured": false, 00:14:01.557 "data_offset": 0, 00:14:01.557 "data_size": 63488 00:14:01.557 }, 00:14:01.557 { 00:14:01.557 "name": "BaseBdev2", 00:14:01.557 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:01.557 "is_configured": true, 00:14:01.557 "data_offset": 2048, 00:14:01.557 "data_size": 63488 00:14:01.557 }, 00:14:01.557 { 00:14:01.557 "name": "BaseBdev3", 00:14:01.557 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:01.557 "is_configured": true, 00:14:01.557 "data_offset": 2048, 00:14:01.557 "data_size": 63488 00:14:01.557 } 00:14:01.557 ] 00:14:01.557 }' 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.557 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.125 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:02.125 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.125 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.125 [2024-09-29 16:52:23.499331] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:02.125 [2024-09-29 16:52:23.499425] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:02.125 [2024-09-29 16:52:23.499464] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:14:02.125 [2024-09-29 16:52:23.499491] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:02.125 [2024-09-29 16:52:23.499991] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:02.125 [2024-09-29 16:52:23.500049] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:02.125 [2024-09-29 16:52:23.500179] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:02.125 [2024-09-29 16:52:23.500218] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:02.125 [2024-09-29 16:52:23.500273] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:02.125 [2024-09-29 16:52:23.500339] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:02.125 [2024-09-29 16:52:23.503532] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043ef0 00:14:02.125 spare 00:14:02.125 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.125 16:52:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:02.125 [2024-09-29 16:52:23.505704] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.065 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:03.065 "name": "raid_bdev1", 00:14:03.065 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:03.065 "strip_size_kb": 64, 00:14:03.065 "state": "online", 00:14:03.065 "raid_level": "raid5f", 00:14:03.065 "superblock": true, 00:14:03.066 "num_base_bdevs": 3, 00:14:03.066 "num_base_bdevs_discovered": 3, 00:14:03.066 "num_base_bdevs_operational": 3, 00:14:03.066 "process": { 00:14:03.066 "type": "rebuild", 00:14:03.066 "target": "spare", 00:14:03.066 "progress": { 00:14:03.066 "blocks": 20480, 00:14:03.066 "percent": 16 00:14:03.066 } 00:14:03.066 }, 00:14:03.066 "base_bdevs_list": [ 00:14:03.066 { 00:14:03.066 "name": "spare", 00:14:03.066 "uuid": "edf93764-7e66-55c5-9972-d319596d59b3", 00:14:03.066 "is_configured": true, 00:14:03.066 "data_offset": 2048, 00:14:03.066 "data_size": 63488 00:14:03.066 }, 00:14:03.066 { 00:14:03.066 "name": "BaseBdev2", 00:14:03.066 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:03.066 "is_configured": true, 00:14:03.066 "data_offset": 2048, 00:14:03.066 "data_size": 63488 00:14:03.066 }, 00:14:03.066 { 00:14:03.066 "name": "BaseBdev3", 00:14:03.066 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:03.066 "is_configured": true, 00:14:03.066 "data_offset": 2048, 00:14:03.066 "data_size": 63488 00:14:03.066 } 00:14:03.066 ] 00:14:03.066 }' 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.066 [2024-09-29 16:52:24.670586] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:03.066 [2024-09-29 16:52:24.712487] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:03.066 [2024-09-29 16:52:24.712550] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:03.066 [2024-09-29 16:52:24.712567] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:03.066 [2024-09-29 16:52:24.712578] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.066 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.326 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.326 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:03.326 "name": "raid_bdev1", 00:14:03.326 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:03.326 "strip_size_kb": 64, 00:14:03.326 "state": "online", 00:14:03.326 "raid_level": "raid5f", 00:14:03.326 "superblock": true, 00:14:03.326 "num_base_bdevs": 3, 00:14:03.326 "num_base_bdevs_discovered": 2, 00:14:03.326 "num_base_bdevs_operational": 2, 00:14:03.326 "base_bdevs_list": [ 00:14:03.326 { 00:14:03.326 "name": null, 00:14:03.326 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.326 "is_configured": false, 00:14:03.326 "data_offset": 0, 00:14:03.326 "data_size": 63488 00:14:03.326 }, 00:14:03.326 { 00:14:03.326 "name": "BaseBdev2", 00:14:03.326 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:03.326 "is_configured": true, 00:14:03.326 "data_offset": 2048, 00:14:03.326 "data_size": 63488 00:14:03.326 }, 00:14:03.326 { 00:14:03.326 "name": "BaseBdev3", 00:14:03.326 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:03.326 "is_configured": true, 00:14:03.326 "data_offset": 2048, 00:14:03.326 "data_size": 63488 00:14:03.326 } 00:14:03.326 ] 00:14:03.326 }' 00:14:03.326 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:03.326 16:52:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:03.586 "name": "raid_bdev1", 00:14:03.586 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:03.586 "strip_size_kb": 64, 00:14:03.586 "state": "online", 00:14:03.586 "raid_level": "raid5f", 00:14:03.586 "superblock": true, 00:14:03.586 "num_base_bdevs": 3, 00:14:03.586 "num_base_bdevs_discovered": 2, 00:14:03.586 "num_base_bdevs_operational": 2, 00:14:03.586 "base_bdevs_list": [ 00:14:03.586 { 00:14:03.586 "name": null, 00:14:03.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.586 "is_configured": false, 00:14:03.586 "data_offset": 0, 00:14:03.586 "data_size": 63488 00:14:03.586 }, 00:14:03.586 { 00:14:03.586 "name": "BaseBdev2", 00:14:03.586 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:03.586 "is_configured": true, 00:14:03.586 "data_offset": 2048, 00:14:03.586 "data_size": 63488 00:14:03.586 }, 00:14:03.586 { 00:14:03.586 "name": "BaseBdev3", 00:14:03.586 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:03.586 "is_configured": true, 00:14:03.586 "data_offset": 2048, 00:14:03.586 "data_size": 63488 00:14:03.586 } 00:14:03.586 ] 00:14:03.586 }' 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:03.586 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:03.846 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:03.846 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:03.846 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.846 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.846 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.846 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:03.847 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.847 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.847 [2024-09-29 16:52:25.308506] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:03.847 [2024-09-29 16:52:25.308559] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:03.847 [2024-09-29 16:52:25.308584] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:03.847 [2024-09-29 16:52:25.308597] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:03.847 [2024-09-29 16:52:25.309005] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:03.847 [2024-09-29 16:52:25.309033] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:03.847 [2024-09-29 16:52:25.309098] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:03.847 [2024-09-29 16:52:25.309113] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:03.847 [2024-09-29 16:52:25.309120] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:03.847 [2024-09-29 16:52:25.309131] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:03.847 BaseBdev1 00:14:03.847 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.847 16:52:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.793 "name": "raid_bdev1", 00:14:04.793 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:04.793 "strip_size_kb": 64, 00:14:04.793 "state": "online", 00:14:04.793 "raid_level": "raid5f", 00:14:04.793 "superblock": true, 00:14:04.793 "num_base_bdevs": 3, 00:14:04.793 "num_base_bdevs_discovered": 2, 00:14:04.793 "num_base_bdevs_operational": 2, 00:14:04.793 "base_bdevs_list": [ 00:14:04.793 { 00:14:04.793 "name": null, 00:14:04.793 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:04.793 "is_configured": false, 00:14:04.793 "data_offset": 0, 00:14:04.793 "data_size": 63488 00:14:04.793 }, 00:14:04.793 { 00:14:04.793 "name": "BaseBdev2", 00:14:04.793 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:04.793 "is_configured": true, 00:14:04.793 "data_offset": 2048, 00:14:04.793 "data_size": 63488 00:14:04.793 }, 00:14:04.793 { 00:14:04.793 "name": "BaseBdev3", 00:14:04.793 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:04.793 "is_configured": true, 00:14:04.793 "data_offset": 2048, 00:14:04.793 "data_size": 63488 00:14:04.793 } 00:14:04.793 ] 00:14:04.793 }' 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.793 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.366 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:05.366 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:05.366 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:05.366 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:05.367 "name": "raid_bdev1", 00:14:05.367 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:05.367 "strip_size_kb": 64, 00:14:05.367 "state": "online", 00:14:05.367 "raid_level": "raid5f", 00:14:05.367 "superblock": true, 00:14:05.367 "num_base_bdevs": 3, 00:14:05.367 "num_base_bdevs_discovered": 2, 00:14:05.367 "num_base_bdevs_operational": 2, 00:14:05.367 "base_bdevs_list": [ 00:14:05.367 { 00:14:05.367 "name": null, 00:14:05.367 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:05.367 "is_configured": false, 00:14:05.367 "data_offset": 0, 00:14:05.367 "data_size": 63488 00:14:05.367 }, 00:14:05.367 { 00:14:05.367 "name": "BaseBdev2", 00:14:05.367 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:05.367 "is_configured": true, 00:14:05.367 "data_offset": 2048, 00:14:05.367 "data_size": 63488 00:14:05.367 }, 00:14:05.367 { 00:14:05.367 "name": "BaseBdev3", 00:14:05.367 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:05.367 "is_configured": true, 00:14:05.367 "data_offset": 2048, 00:14:05.367 "data_size": 63488 00:14:05.367 } 00:14:05.367 ] 00:14:05.367 }' 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.367 [2024-09-29 16:52:26.926029] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:05.367 [2024-09-29 16:52:26.926178] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:05.367 [2024-09-29 16:52:26.926190] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:05.367 request: 00:14:05.367 { 00:14:05.367 "base_bdev": "BaseBdev1", 00:14:05.367 "raid_bdev": "raid_bdev1", 00:14:05.367 "method": "bdev_raid_add_base_bdev", 00:14:05.367 "req_id": 1 00:14:05.367 } 00:14:05.367 Got JSON-RPC error response 00:14:05.367 response: 00:14:05.367 { 00:14:05.367 "code": -22, 00:14:05.367 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:05.367 } 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:05.367 16:52:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.349 "name": "raid_bdev1", 00:14:06.349 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:06.349 "strip_size_kb": 64, 00:14:06.349 "state": "online", 00:14:06.349 "raid_level": "raid5f", 00:14:06.349 "superblock": true, 00:14:06.349 "num_base_bdevs": 3, 00:14:06.349 "num_base_bdevs_discovered": 2, 00:14:06.349 "num_base_bdevs_operational": 2, 00:14:06.349 "base_bdevs_list": [ 00:14:06.349 { 00:14:06.349 "name": null, 00:14:06.349 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.349 "is_configured": false, 00:14:06.349 "data_offset": 0, 00:14:06.349 "data_size": 63488 00:14:06.349 }, 00:14:06.349 { 00:14:06.349 "name": "BaseBdev2", 00:14:06.349 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:06.349 "is_configured": true, 00:14:06.349 "data_offset": 2048, 00:14:06.349 "data_size": 63488 00:14:06.349 }, 00:14:06.349 { 00:14:06.349 "name": "BaseBdev3", 00:14:06.349 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:06.349 "is_configured": true, 00:14:06.349 "data_offset": 2048, 00:14:06.349 "data_size": 63488 00:14:06.349 } 00:14:06.349 ] 00:14:06.349 }' 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.349 16:52:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.918 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:06.918 "name": "raid_bdev1", 00:14:06.918 "uuid": "f57c8682-4b4b-4072-ba38-e612228a7795", 00:14:06.918 "strip_size_kb": 64, 00:14:06.918 "state": "online", 00:14:06.918 "raid_level": "raid5f", 00:14:06.918 "superblock": true, 00:14:06.918 "num_base_bdevs": 3, 00:14:06.919 "num_base_bdevs_discovered": 2, 00:14:06.919 "num_base_bdevs_operational": 2, 00:14:06.919 "base_bdevs_list": [ 00:14:06.919 { 00:14:06.919 "name": null, 00:14:06.919 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:06.919 "is_configured": false, 00:14:06.919 "data_offset": 0, 00:14:06.919 "data_size": 63488 00:14:06.919 }, 00:14:06.919 { 00:14:06.919 "name": "BaseBdev2", 00:14:06.919 "uuid": "9c65defe-823f-51ec-ad19-e028cbf097e5", 00:14:06.919 "is_configured": true, 00:14:06.919 "data_offset": 2048, 00:14:06.919 "data_size": 63488 00:14:06.919 }, 00:14:06.919 { 00:14:06.919 "name": "BaseBdev3", 00:14:06.919 "uuid": "bff9ee14-dd32-5b81-93c6-9c161eeea289", 00:14:06.919 "is_configured": true, 00:14:06.919 "data_offset": 2048, 00:14:06.919 "data_size": 63488 00:14:06.919 } 00:14:06.919 ] 00:14:06.919 }' 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92159 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 92159 ']' 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 92159 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:06.919 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 92159 00:14:07.179 killing process with pid 92159 00:14:07.179 Received shutdown signal, test time was about 60.000000 seconds 00:14:07.179 00:14:07.179 Latency(us) 00:14:07.179 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:07.179 =================================================================================================================== 00:14:07.179 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:07.179 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:07.179 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:07.179 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 92159' 00:14:07.179 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 92159 00:14:07.179 [2024-09-29 16:52:28.612951] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:07.179 [2024-09-29 16:52:28.613066] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:07.179 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 92159 00:14:07.179 [2024-09-29 16:52:28.613129] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:07.179 [2024-09-29 16:52:28.613138] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:14:07.179 [2024-09-29 16:52:28.654132] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:07.439 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:07.439 00:14:07.439 real 0m21.666s 00:14:07.439 user 0m28.138s 00:14:07.439 sys 0m2.870s 00:14:07.439 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:07.439 ************************************ 00:14:07.439 END TEST raid5f_rebuild_test_sb 00:14:07.439 ************************************ 00:14:07.439 16:52:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.439 16:52:28 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:07.439 16:52:28 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:14:07.439 16:52:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:14:07.439 16:52:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:07.439 16:52:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:07.439 ************************************ 00:14:07.439 START TEST raid5f_state_function_test 00:14:07.439 ************************************ 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 false 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=92894 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 92894' 00:14:07.439 Process raid pid: 92894 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 92894 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 92894 ']' 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:07.439 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:07.439 16:52:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.440 [2024-09-29 16:52:29.071651] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:14:07.440 [2024-09-29 16:52:29.071865] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:07.699 [2024-09-29 16:52:29.219965] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:07.699 [2024-09-29 16:52:29.265999] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:07.699 [2024-09-29 16:52:29.308906] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:07.699 [2024-09-29 16:52:29.309019] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:08.268 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:08.268 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:14:08.268 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:08.268 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.268 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.268 [2024-09-29 16:52:29.902789] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:08.268 [2024-09-29 16:52:29.902832] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:08.268 [2024-09-29 16:52:29.902851] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:08.268 [2024-09-29 16:52:29.902862] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:08.268 [2024-09-29 16:52:29.902868] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:08.268 [2024-09-29 16:52:29.902879] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:08.268 [2024-09-29 16:52:29.902885] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:08.268 [2024-09-29 16:52:29.902893] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:08.268 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.269 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.527 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:08.527 "name": "Existed_Raid", 00:14:08.527 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.527 "strip_size_kb": 64, 00:14:08.527 "state": "configuring", 00:14:08.527 "raid_level": "raid5f", 00:14:08.527 "superblock": false, 00:14:08.527 "num_base_bdevs": 4, 00:14:08.527 "num_base_bdevs_discovered": 0, 00:14:08.527 "num_base_bdevs_operational": 4, 00:14:08.527 "base_bdevs_list": [ 00:14:08.527 { 00:14:08.528 "name": "BaseBdev1", 00:14:08.528 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.528 "is_configured": false, 00:14:08.528 "data_offset": 0, 00:14:08.528 "data_size": 0 00:14:08.528 }, 00:14:08.528 { 00:14:08.528 "name": "BaseBdev2", 00:14:08.528 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.528 "is_configured": false, 00:14:08.528 "data_offset": 0, 00:14:08.528 "data_size": 0 00:14:08.528 }, 00:14:08.528 { 00:14:08.528 "name": "BaseBdev3", 00:14:08.528 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.528 "is_configured": false, 00:14:08.528 "data_offset": 0, 00:14:08.528 "data_size": 0 00:14:08.528 }, 00:14:08.528 { 00:14:08.528 "name": "BaseBdev4", 00:14:08.528 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:08.528 "is_configured": false, 00:14:08.528 "data_offset": 0, 00:14:08.528 "data_size": 0 00:14:08.528 } 00:14:08.528 ] 00:14:08.528 }' 00:14:08.528 16:52:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:08.528 16:52:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.788 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:08.788 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.788 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.788 [2024-09-29 16:52:30.393810] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:08.788 [2024-09-29 16:52:30.393885] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:08.788 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.788 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:08.788 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.788 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.789 [2024-09-29 16:52:30.401826] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:08.789 [2024-09-29 16:52:30.401916] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:08.789 [2024-09-29 16:52:30.401942] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:08.789 [2024-09-29 16:52:30.401964] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:08.789 [2024-09-29 16:52:30.401981] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:08.789 [2024-09-29 16:52:30.402002] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:08.789 [2024-09-29 16:52:30.402019] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:08.789 [2024-09-29 16:52:30.402061] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.789 [2024-09-29 16:52:30.418749] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:08.789 BaseBdev1 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.789 [ 00:14:08.789 { 00:14:08.789 "name": "BaseBdev1", 00:14:08.789 "aliases": [ 00:14:08.789 "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202" 00:14:08.789 ], 00:14:08.789 "product_name": "Malloc disk", 00:14:08.789 "block_size": 512, 00:14:08.789 "num_blocks": 65536, 00:14:08.789 "uuid": "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202", 00:14:08.789 "assigned_rate_limits": { 00:14:08.789 "rw_ios_per_sec": 0, 00:14:08.789 "rw_mbytes_per_sec": 0, 00:14:08.789 "r_mbytes_per_sec": 0, 00:14:08.789 "w_mbytes_per_sec": 0 00:14:08.789 }, 00:14:08.789 "claimed": true, 00:14:08.789 "claim_type": "exclusive_write", 00:14:08.789 "zoned": false, 00:14:08.789 "supported_io_types": { 00:14:08.789 "read": true, 00:14:08.789 "write": true, 00:14:08.789 "unmap": true, 00:14:08.789 "flush": true, 00:14:08.789 "reset": true, 00:14:08.789 "nvme_admin": false, 00:14:08.789 "nvme_io": false, 00:14:08.789 "nvme_io_md": false, 00:14:08.789 "write_zeroes": true, 00:14:08.789 "zcopy": true, 00:14:08.789 "get_zone_info": false, 00:14:08.789 "zone_management": false, 00:14:08.789 "zone_append": false, 00:14:08.789 "compare": false, 00:14:08.789 "compare_and_write": false, 00:14:08.789 "abort": true, 00:14:08.789 "seek_hole": false, 00:14:08.789 "seek_data": false, 00:14:08.789 "copy": true, 00:14:08.789 "nvme_iov_md": false 00:14:08.789 }, 00:14:08.789 "memory_domains": [ 00:14:08.789 { 00:14:08.789 "dma_device_id": "system", 00:14:08.789 "dma_device_type": 1 00:14:08.789 }, 00:14:08.789 { 00:14:08.789 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:08.789 "dma_device_type": 2 00:14:08.789 } 00:14:08.789 ], 00:14:08.789 "driver_specific": {} 00:14:08.789 } 00:14:08.789 ] 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:08.789 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.049 "name": "Existed_Raid", 00:14:09.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.049 "strip_size_kb": 64, 00:14:09.049 "state": "configuring", 00:14:09.049 "raid_level": "raid5f", 00:14:09.049 "superblock": false, 00:14:09.049 "num_base_bdevs": 4, 00:14:09.049 "num_base_bdevs_discovered": 1, 00:14:09.049 "num_base_bdevs_operational": 4, 00:14:09.049 "base_bdevs_list": [ 00:14:09.049 { 00:14:09.049 "name": "BaseBdev1", 00:14:09.049 "uuid": "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202", 00:14:09.049 "is_configured": true, 00:14:09.049 "data_offset": 0, 00:14:09.049 "data_size": 65536 00:14:09.049 }, 00:14:09.049 { 00:14:09.049 "name": "BaseBdev2", 00:14:09.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.049 "is_configured": false, 00:14:09.049 "data_offset": 0, 00:14:09.049 "data_size": 0 00:14:09.049 }, 00:14:09.049 { 00:14:09.049 "name": "BaseBdev3", 00:14:09.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.049 "is_configured": false, 00:14:09.049 "data_offset": 0, 00:14:09.049 "data_size": 0 00:14:09.049 }, 00:14:09.049 { 00:14:09.049 "name": "BaseBdev4", 00:14:09.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.049 "is_configured": false, 00:14:09.049 "data_offset": 0, 00:14:09.049 "data_size": 0 00:14:09.049 } 00:14:09.049 ] 00:14:09.049 }' 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.049 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.309 [2024-09-29 16:52:30.929886] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:09.309 [2024-09-29 16:52:30.929983] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.309 [2024-09-29 16:52:30.937921] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:09.309 [2024-09-29 16:52:30.939839] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:09.309 [2024-09-29 16:52:30.939880] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:09.309 [2024-09-29 16:52:30.939889] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:09.309 [2024-09-29 16:52:30.939898] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:09.309 [2024-09-29 16:52:30.939904] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:09.309 [2024-09-29 16:52:30.939912] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:09.309 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.310 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.570 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.570 "name": "Existed_Raid", 00:14:09.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.570 "strip_size_kb": 64, 00:14:09.570 "state": "configuring", 00:14:09.570 "raid_level": "raid5f", 00:14:09.570 "superblock": false, 00:14:09.570 "num_base_bdevs": 4, 00:14:09.570 "num_base_bdevs_discovered": 1, 00:14:09.570 "num_base_bdevs_operational": 4, 00:14:09.570 "base_bdevs_list": [ 00:14:09.570 { 00:14:09.570 "name": "BaseBdev1", 00:14:09.570 "uuid": "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202", 00:14:09.570 "is_configured": true, 00:14:09.570 "data_offset": 0, 00:14:09.570 "data_size": 65536 00:14:09.570 }, 00:14:09.570 { 00:14:09.570 "name": "BaseBdev2", 00:14:09.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.570 "is_configured": false, 00:14:09.570 "data_offset": 0, 00:14:09.570 "data_size": 0 00:14:09.570 }, 00:14:09.570 { 00:14:09.570 "name": "BaseBdev3", 00:14:09.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.570 "is_configured": false, 00:14:09.570 "data_offset": 0, 00:14:09.570 "data_size": 0 00:14:09.570 }, 00:14:09.570 { 00:14:09.570 "name": "BaseBdev4", 00:14:09.570 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:09.570 "is_configured": false, 00:14:09.570 "data_offset": 0, 00:14:09.570 "data_size": 0 00:14:09.570 } 00:14:09.570 ] 00:14:09.570 }' 00:14:09.570 16:52:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.570 16:52:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.830 [2024-09-29 16:52:31.407376] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:09.830 BaseBdev2 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.830 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.830 [ 00:14:09.830 { 00:14:09.830 "name": "BaseBdev2", 00:14:09.830 "aliases": [ 00:14:09.830 "f990c01e-ce08-405b-93e4-dd894e03023d" 00:14:09.830 ], 00:14:09.830 "product_name": "Malloc disk", 00:14:09.830 "block_size": 512, 00:14:09.830 "num_blocks": 65536, 00:14:09.830 "uuid": "f990c01e-ce08-405b-93e4-dd894e03023d", 00:14:09.830 "assigned_rate_limits": { 00:14:09.830 "rw_ios_per_sec": 0, 00:14:09.830 "rw_mbytes_per_sec": 0, 00:14:09.830 "r_mbytes_per_sec": 0, 00:14:09.830 "w_mbytes_per_sec": 0 00:14:09.830 }, 00:14:09.830 "claimed": true, 00:14:09.830 "claim_type": "exclusive_write", 00:14:09.830 "zoned": false, 00:14:09.830 "supported_io_types": { 00:14:09.831 "read": true, 00:14:09.831 "write": true, 00:14:09.831 "unmap": true, 00:14:09.831 "flush": true, 00:14:09.831 "reset": true, 00:14:09.831 "nvme_admin": false, 00:14:09.831 "nvme_io": false, 00:14:09.831 "nvme_io_md": false, 00:14:09.831 "write_zeroes": true, 00:14:09.831 "zcopy": true, 00:14:09.831 "get_zone_info": false, 00:14:09.831 "zone_management": false, 00:14:09.831 "zone_append": false, 00:14:09.831 "compare": false, 00:14:09.831 "compare_and_write": false, 00:14:09.831 "abort": true, 00:14:09.831 "seek_hole": false, 00:14:09.831 "seek_data": false, 00:14:09.831 "copy": true, 00:14:09.831 "nvme_iov_md": false 00:14:09.831 }, 00:14:09.831 "memory_domains": [ 00:14:09.831 { 00:14:09.831 "dma_device_id": "system", 00:14:09.831 "dma_device_type": 1 00:14:09.831 }, 00:14:09.831 { 00:14:09.831 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:09.831 "dma_device_type": 2 00:14:09.831 } 00:14:09.831 ], 00:14:09.831 "driver_specific": {} 00:14:09.831 } 00:14:09.831 ] 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.831 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.091 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.091 "name": "Existed_Raid", 00:14:10.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.091 "strip_size_kb": 64, 00:14:10.091 "state": "configuring", 00:14:10.091 "raid_level": "raid5f", 00:14:10.091 "superblock": false, 00:14:10.091 "num_base_bdevs": 4, 00:14:10.091 "num_base_bdevs_discovered": 2, 00:14:10.091 "num_base_bdevs_operational": 4, 00:14:10.091 "base_bdevs_list": [ 00:14:10.091 { 00:14:10.091 "name": "BaseBdev1", 00:14:10.091 "uuid": "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202", 00:14:10.091 "is_configured": true, 00:14:10.091 "data_offset": 0, 00:14:10.091 "data_size": 65536 00:14:10.091 }, 00:14:10.091 { 00:14:10.091 "name": "BaseBdev2", 00:14:10.091 "uuid": "f990c01e-ce08-405b-93e4-dd894e03023d", 00:14:10.091 "is_configured": true, 00:14:10.091 "data_offset": 0, 00:14:10.091 "data_size": 65536 00:14:10.091 }, 00:14:10.091 { 00:14:10.091 "name": "BaseBdev3", 00:14:10.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.091 "is_configured": false, 00:14:10.091 "data_offset": 0, 00:14:10.091 "data_size": 0 00:14:10.091 }, 00:14:10.091 { 00:14:10.091 "name": "BaseBdev4", 00:14:10.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.091 "is_configured": false, 00:14:10.091 "data_offset": 0, 00:14:10.091 "data_size": 0 00:14:10.091 } 00:14:10.091 ] 00:14:10.091 }' 00:14:10.091 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.091 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.351 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:10.351 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.351 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.351 BaseBdev3 00:14:10.351 [2024-09-29 16:52:31.929784] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:10.351 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.351 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.352 [ 00:14:10.352 { 00:14:10.352 "name": "BaseBdev3", 00:14:10.352 "aliases": [ 00:14:10.352 "efa3770e-825b-433d-abcf-03d8402578ac" 00:14:10.352 ], 00:14:10.352 "product_name": "Malloc disk", 00:14:10.352 "block_size": 512, 00:14:10.352 "num_blocks": 65536, 00:14:10.352 "uuid": "efa3770e-825b-433d-abcf-03d8402578ac", 00:14:10.352 "assigned_rate_limits": { 00:14:10.352 "rw_ios_per_sec": 0, 00:14:10.352 "rw_mbytes_per_sec": 0, 00:14:10.352 "r_mbytes_per_sec": 0, 00:14:10.352 "w_mbytes_per_sec": 0 00:14:10.352 }, 00:14:10.352 "claimed": true, 00:14:10.352 "claim_type": "exclusive_write", 00:14:10.352 "zoned": false, 00:14:10.352 "supported_io_types": { 00:14:10.352 "read": true, 00:14:10.352 "write": true, 00:14:10.352 "unmap": true, 00:14:10.352 "flush": true, 00:14:10.352 "reset": true, 00:14:10.352 "nvme_admin": false, 00:14:10.352 "nvme_io": false, 00:14:10.352 "nvme_io_md": false, 00:14:10.352 "write_zeroes": true, 00:14:10.352 "zcopy": true, 00:14:10.352 "get_zone_info": false, 00:14:10.352 "zone_management": false, 00:14:10.352 "zone_append": false, 00:14:10.352 "compare": false, 00:14:10.352 "compare_and_write": false, 00:14:10.352 "abort": true, 00:14:10.352 "seek_hole": false, 00:14:10.352 "seek_data": false, 00:14:10.352 "copy": true, 00:14:10.352 "nvme_iov_md": false 00:14:10.352 }, 00:14:10.352 "memory_domains": [ 00:14:10.352 { 00:14:10.352 "dma_device_id": "system", 00:14:10.352 "dma_device_type": 1 00:14:10.352 }, 00:14:10.352 { 00:14:10.352 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:10.352 "dma_device_type": 2 00:14:10.352 } 00:14:10.352 ], 00:14:10.352 "driver_specific": {} 00:14:10.352 } 00:14:10.352 ] 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.352 16:52:31 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.352 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.352 "name": "Existed_Raid", 00:14:10.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.352 "strip_size_kb": 64, 00:14:10.352 "state": "configuring", 00:14:10.352 "raid_level": "raid5f", 00:14:10.352 "superblock": false, 00:14:10.352 "num_base_bdevs": 4, 00:14:10.352 "num_base_bdevs_discovered": 3, 00:14:10.352 "num_base_bdevs_operational": 4, 00:14:10.352 "base_bdevs_list": [ 00:14:10.352 { 00:14:10.352 "name": "BaseBdev1", 00:14:10.352 "uuid": "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202", 00:14:10.352 "is_configured": true, 00:14:10.352 "data_offset": 0, 00:14:10.352 "data_size": 65536 00:14:10.352 }, 00:14:10.352 { 00:14:10.352 "name": "BaseBdev2", 00:14:10.352 "uuid": "f990c01e-ce08-405b-93e4-dd894e03023d", 00:14:10.352 "is_configured": true, 00:14:10.352 "data_offset": 0, 00:14:10.352 "data_size": 65536 00:14:10.352 }, 00:14:10.352 { 00:14:10.352 "name": "BaseBdev3", 00:14:10.352 "uuid": "efa3770e-825b-433d-abcf-03d8402578ac", 00:14:10.352 "is_configured": true, 00:14:10.352 "data_offset": 0, 00:14:10.352 "data_size": 65536 00:14:10.352 }, 00:14:10.352 { 00:14:10.352 "name": "BaseBdev4", 00:14:10.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:10.352 "is_configured": false, 00:14:10.352 "data_offset": 0, 00:14:10.352 "data_size": 0 00:14:10.352 } 00:14:10.352 ] 00:14:10.352 }' 00:14:10.352 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.352 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.922 BaseBdev4 00:14:10.922 [2024-09-29 16:52:32.439891] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:10.922 [2024-09-29 16:52:32.439948] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:10.922 [2024-09-29 16:52:32.439956] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:10.922 [2024-09-29 16:52:32.440213] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:10.922 [2024-09-29 16:52:32.440677] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:10.922 [2024-09-29 16:52:32.440696] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:10.922 [2024-09-29 16:52:32.440928] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.922 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.922 [ 00:14:10.922 { 00:14:10.922 "name": "BaseBdev4", 00:14:10.922 "aliases": [ 00:14:10.922 "d4425bb5-9238-441f-9b20-9262f45a552c" 00:14:10.922 ], 00:14:10.922 "product_name": "Malloc disk", 00:14:10.922 "block_size": 512, 00:14:10.922 "num_blocks": 65536, 00:14:10.922 "uuid": "d4425bb5-9238-441f-9b20-9262f45a552c", 00:14:10.922 "assigned_rate_limits": { 00:14:10.922 "rw_ios_per_sec": 0, 00:14:10.922 "rw_mbytes_per_sec": 0, 00:14:10.922 "r_mbytes_per_sec": 0, 00:14:10.922 "w_mbytes_per_sec": 0 00:14:10.922 }, 00:14:10.922 "claimed": true, 00:14:10.922 "claim_type": "exclusive_write", 00:14:10.922 "zoned": false, 00:14:10.922 "supported_io_types": { 00:14:10.922 "read": true, 00:14:10.922 "write": true, 00:14:10.922 "unmap": true, 00:14:10.922 "flush": true, 00:14:10.922 "reset": true, 00:14:10.922 "nvme_admin": false, 00:14:10.922 "nvme_io": false, 00:14:10.922 "nvme_io_md": false, 00:14:10.922 "write_zeroes": true, 00:14:10.922 "zcopy": true, 00:14:10.922 "get_zone_info": false, 00:14:10.922 "zone_management": false, 00:14:10.922 "zone_append": false, 00:14:10.922 "compare": false, 00:14:10.922 "compare_and_write": false, 00:14:10.922 "abort": true, 00:14:10.922 "seek_hole": false, 00:14:10.922 "seek_data": false, 00:14:10.922 "copy": true, 00:14:10.922 "nvme_iov_md": false 00:14:10.922 }, 00:14:10.922 "memory_domains": [ 00:14:10.922 { 00:14:10.922 "dma_device_id": "system", 00:14:10.923 "dma_device_type": 1 00:14:10.923 }, 00:14:10.923 { 00:14:10.923 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:10.923 "dma_device_type": 2 00:14:10.923 } 00:14:10.923 ], 00:14:10.923 "driver_specific": {} 00:14:10.923 } 00:14:10.923 ] 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.923 "name": "Existed_Raid", 00:14:10.923 "uuid": "7cbef313-8e06-4607-9414-68453425560b", 00:14:10.923 "strip_size_kb": 64, 00:14:10.923 "state": "online", 00:14:10.923 "raid_level": "raid5f", 00:14:10.923 "superblock": false, 00:14:10.923 "num_base_bdevs": 4, 00:14:10.923 "num_base_bdevs_discovered": 4, 00:14:10.923 "num_base_bdevs_operational": 4, 00:14:10.923 "base_bdevs_list": [ 00:14:10.923 { 00:14:10.923 "name": "BaseBdev1", 00:14:10.923 "uuid": "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202", 00:14:10.923 "is_configured": true, 00:14:10.923 "data_offset": 0, 00:14:10.923 "data_size": 65536 00:14:10.923 }, 00:14:10.923 { 00:14:10.923 "name": "BaseBdev2", 00:14:10.923 "uuid": "f990c01e-ce08-405b-93e4-dd894e03023d", 00:14:10.923 "is_configured": true, 00:14:10.923 "data_offset": 0, 00:14:10.923 "data_size": 65536 00:14:10.923 }, 00:14:10.923 { 00:14:10.923 "name": "BaseBdev3", 00:14:10.923 "uuid": "efa3770e-825b-433d-abcf-03d8402578ac", 00:14:10.923 "is_configured": true, 00:14:10.923 "data_offset": 0, 00:14:10.923 "data_size": 65536 00:14:10.923 }, 00:14:10.923 { 00:14:10.923 "name": "BaseBdev4", 00:14:10.923 "uuid": "d4425bb5-9238-441f-9b20-9262f45a552c", 00:14:10.923 "is_configured": true, 00:14:10.923 "data_offset": 0, 00:14:10.923 "data_size": 65536 00:14:10.923 } 00:14:10.923 ] 00:14:10.923 }' 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.923 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.493 [2024-09-29 16:52:32.927373] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.493 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:11.493 "name": "Existed_Raid", 00:14:11.493 "aliases": [ 00:14:11.493 "7cbef313-8e06-4607-9414-68453425560b" 00:14:11.493 ], 00:14:11.493 "product_name": "Raid Volume", 00:14:11.493 "block_size": 512, 00:14:11.493 "num_blocks": 196608, 00:14:11.493 "uuid": "7cbef313-8e06-4607-9414-68453425560b", 00:14:11.493 "assigned_rate_limits": { 00:14:11.493 "rw_ios_per_sec": 0, 00:14:11.493 "rw_mbytes_per_sec": 0, 00:14:11.493 "r_mbytes_per_sec": 0, 00:14:11.493 "w_mbytes_per_sec": 0 00:14:11.493 }, 00:14:11.493 "claimed": false, 00:14:11.493 "zoned": false, 00:14:11.493 "supported_io_types": { 00:14:11.493 "read": true, 00:14:11.493 "write": true, 00:14:11.493 "unmap": false, 00:14:11.493 "flush": false, 00:14:11.493 "reset": true, 00:14:11.493 "nvme_admin": false, 00:14:11.493 "nvme_io": false, 00:14:11.493 "nvme_io_md": false, 00:14:11.493 "write_zeroes": true, 00:14:11.493 "zcopy": false, 00:14:11.493 "get_zone_info": false, 00:14:11.493 "zone_management": false, 00:14:11.493 "zone_append": false, 00:14:11.493 "compare": false, 00:14:11.493 "compare_and_write": false, 00:14:11.493 "abort": false, 00:14:11.493 "seek_hole": false, 00:14:11.493 "seek_data": false, 00:14:11.493 "copy": false, 00:14:11.493 "nvme_iov_md": false 00:14:11.493 }, 00:14:11.493 "driver_specific": { 00:14:11.493 "raid": { 00:14:11.493 "uuid": "7cbef313-8e06-4607-9414-68453425560b", 00:14:11.493 "strip_size_kb": 64, 00:14:11.493 "state": "online", 00:14:11.493 "raid_level": "raid5f", 00:14:11.493 "superblock": false, 00:14:11.493 "num_base_bdevs": 4, 00:14:11.493 "num_base_bdevs_discovered": 4, 00:14:11.493 "num_base_bdevs_operational": 4, 00:14:11.493 "base_bdevs_list": [ 00:14:11.493 { 00:14:11.493 "name": "BaseBdev1", 00:14:11.493 "uuid": "d7a58c2c-3d84-41c8-99bc-a9b16f8d0202", 00:14:11.493 "is_configured": true, 00:14:11.493 "data_offset": 0, 00:14:11.493 "data_size": 65536 00:14:11.493 }, 00:14:11.493 { 00:14:11.493 "name": "BaseBdev2", 00:14:11.493 "uuid": "f990c01e-ce08-405b-93e4-dd894e03023d", 00:14:11.494 "is_configured": true, 00:14:11.494 "data_offset": 0, 00:14:11.494 "data_size": 65536 00:14:11.494 }, 00:14:11.494 { 00:14:11.494 "name": "BaseBdev3", 00:14:11.494 "uuid": "efa3770e-825b-433d-abcf-03d8402578ac", 00:14:11.494 "is_configured": true, 00:14:11.494 "data_offset": 0, 00:14:11.494 "data_size": 65536 00:14:11.494 }, 00:14:11.494 { 00:14:11.494 "name": "BaseBdev4", 00:14:11.494 "uuid": "d4425bb5-9238-441f-9b20-9262f45a552c", 00:14:11.494 "is_configured": true, 00:14:11.494 "data_offset": 0, 00:14:11.494 "data_size": 65536 00:14:11.494 } 00:14:11.494 ] 00:14:11.494 } 00:14:11.494 } 00:14:11.494 }' 00:14:11.494 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:11.494 16:52:32 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:11.494 BaseBdev2 00:14:11.494 BaseBdev3 00:14:11.494 BaseBdev4' 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.494 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.754 [2024-09-29 16:52:33.262847] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.754 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.755 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.755 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.755 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.755 "name": "Existed_Raid", 00:14:11.755 "uuid": "7cbef313-8e06-4607-9414-68453425560b", 00:14:11.755 "strip_size_kb": 64, 00:14:11.755 "state": "online", 00:14:11.755 "raid_level": "raid5f", 00:14:11.755 "superblock": false, 00:14:11.755 "num_base_bdevs": 4, 00:14:11.755 "num_base_bdevs_discovered": 3, 00:14:11.755 "num_base_bdevs_operational": 3, 00:14:11.755 "base_bdevs_list": [ 00:14:11.755 { 00:14:11.755 "name": null, 00:14:11.755 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.755 "is_configured": false, 00:14:11.755 "data_offset": 0, 00:14:11.755 "data_size": 65536 00:14:11.755 }, 00:14:11.755 { 00:14:11.755 "name": "BaseBdev2", 00:14:11.755 "uuid": "f990c01e-ce08-405b-93e4-dd894e03023d", 00:14:11.755 "is_configured": true, 00:14:11.755 "data_offset": 0, 00:14:11.755 "data_size": 65536 00:14:11.755 }, 00:14:11.755 { 00:14:11.755 "name": "BaseBdev3", 00:14:11.755 "uuid": "efa3770e-825b-433d-abcf-03d8402578ac", 00:14:11.755 "is_configured": true, 00:14:11.755 "data_offset": 0, 00:14:11.755 "data_size": 65536 00:14:11.755 }, 00:14:11.755 { 00:14:11.755 "name": "BaseBdev4", 00:14:11.755 "uuid": "d4425bb5-9238-441f-9b20-9262f45a552c", 00:14:11.755 "is_configured": true, 00:14:11.755 "data_offset": 0, 00:14:11.755 "data_size": 65536 00:14:11.755 } 00:14:11.755 ] 00:14:11.755 }' 00:14:11.755 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.755 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 [2024-09-29 16:52:33.757504] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:12.326 [2024-09-29 16:52:33.757636] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:12.326 [2024-09-29 16:52:33.769015] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 [2024-09-29 16:52:33.828915] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 [2024-09-29 16:52:33.899845] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:12.326 [2024-09-29 16:52:33.899887] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 BaseBdev2 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.326 16:52:33 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.587 [ 00:14:12.587 { 00:14:12.587 "name": "BaseBdev2", 00:14:12.587 "aliases": [ 00:14:12.587 "a885c108-dcc3-4c30-be65-20e23f1d7985" 00:14:12.587 ], 00:14:12.587 "product_name": "Malloc disk", 00:14:12.587 "block_size": 512, 00:14:12.587 "num_blocks": 65536, 00:14:12.587 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:12.587 "assigned_rate_limits": { 00:14:12.587 "rw_ios_per_sec": 0, 00:14:12.587 "rw_mbytes_per_sec": 0, 00:14:12.587 "r_mbytes_per_sec": 0, 00:14:12.587 "w_mbytes_per_sec": 0 00:14:12.587 }, 00:14:12.587 "claimed": false, 00:14:12.587 "zoned": false, 00:14:12.587 "supported_io_types": { 00:14:12.587 "read": true, 00:14:12.587 "write": true, 00:14:12.587 "unmap": true, 00:14:12.587 "flush": true, 00:14:12.587 "reset": true, 00:14:12.587 "nvme_admin": false, 00:14:12.587 "nvme_io": false, 00:14:12.587 "nvme_io_md": false, 00:14:12.587 "write_zeroes": true, 00:14:12.587 "zcopy": true, 00:14:12.587 "get_zone_info": false, 00:14:12.587 "zone_management": false, 00:14:12.587 "zone_append": false, 00:14:12.587 "compare": false, 00:14:12.587 "compare_and_write": false, 00:14:12.587 "abort": true, 00:14:12.587 "seek_hole": false, 00:14:12.587 "seek_data": false, 00:14:12.587 "copy": true, 00:14:12.587 "nvme_iov_md": false 00:14:12.587 }, 00:14:12.587 "memory_domains": [ 00:14:12.587 { 00:14:12.587 "dma_device_id": "system", 00:14:12.587 "dma_device_type": 1 00:14:12.587 }, 00:14:12.587 { 00:14:12.587 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:12.587 "dma_device_type": 2 00:14:12.587 } 00:14:12.587 ], 00:14:12.587 "driver_specific": {} 00:14:12.587 } 00:14:12.587 ] 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.587 BaseBdev3 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.587 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.587 [ 00:14:12.587 { 00:14:12.587 "name": "BaseBdev3", 00:14:12.587 "aliases": [ 00:14:12.587 "e40e952b-02ed-4bd4-be62-ff258a75f0b1" 00:14:12.587 ], 00:14:12.587 "product_name": "Malloc disk", 00:14:12.587 "block_size": 512, 00:14:12.588 "num_blocks": 65536, 00:14:12.588 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:12.588 "assigned_rate_limits": { 00:14:12.588 "rw_ios_per_sec": 0, 00:14:12.588 "rw_mbytes_per_sec": 0, 00:14:12.588 "r_mbytes_per_sec": 0, 00:14:12.588 "w_mbytes_per_sec": 0 00:14:12.588 }, 00:14:12.588 "claimed": false, 00:14:12.588 "zoned": false, 00:14:12.588 "supported_io_types": { 00:14:12.588 "read": true, 00:14:12.588 "write": true, 00:14:12.588 "unmap": true, 00:14:12.588 "flush": true, 00:14:12.588 "reset": true, 00:14:12.588 "nvme_admin": false, 00:14:12.588 "nvme_io": false, 00:14:12.588 "nvme_io_md": false, 00:14:12.588 "write_zeroes": true, 00:14:12.588 "zcopy": true, 00:14:12.588 "get_zone_info": false, 00:14:12.588 "zone_management": false, 00:14:12.588 "zone_append": false, 00:14:12.588 "compare": false, 00:14:12.588 "compare_and_write": false, 00:14:12.588 "abort": true, 00:14:12.588 "seek_hole": false, 00:14:12.588 "seek_data": false, 00:14:12.588 "copy": true, 00:14:12.588 "nvme_iov_md": false 00:14:12.588 }, 00:14:12.588 "memory_domains": [ 00:14:12.588 { 00:14:12.588 "dma_device_id": "system", 00:14:12.588 "dma_device_type": 1 00:14:12.588 }, 00:14:12.588 { 00:14:12.588 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:12.588 "dma_device_type": 2 00:14:12.588 } 00:14:12.588 ], 00:14:12.588 "driver_specific": {} 00:14:12.588 } 00:14:12.588 ] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.588 BaseBdev4 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.588 [ 00:14:12.588 { 00:14:12.588 "name": "BaseBdev4", 00:14:12.588 "aliases": [ 00:14:12.588 "86b75c21-daba-4724-87a7-8f1a82e7098a" 00:14:12.588 ], 00:14:12.588 "product_name": "Malloc disk", 00:14:12.588 "block_size": 512, 00:14:12.588 "num_blocks": 65536, 00:14:12.588 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:12.588 "assigned_rate_limits": { 00:14:12.588 "rw_ios_per_sec": 0, 00:14:12.588 "rw_mbytes_per_sec": 0, 00:14:12.588 "r_mbytes_per_sec": 0, 00:14:12.588 "w_mbytes_per_sec": 0 00:14:12.588 }, 00:14:12.588 "claimed": false, 00:14:12.588 "zoned": false, 00:14:12.588 "supported_io_types": { 00:14:12.588 "read": true, 00:14:12.588 "write": true, 00:14:12.588 "unmap": true, 00:14:12.588 "flush": true, 00:14:12.588 "reset": true, 00:14:12.588 "nvme_admin": false, 00:14:12.588 "nvme_io": false, 00:14:12.588 "nvme_io_md": false, 00:14:12.588 "write_zeroes": true, 00:14:12.588 "zcopy": true, 00:14:12.588 "get_zone_info": false, 00:14:12.588 "zone_management": false, 00:14:12.588 "zone_append": false, 00:14:12.588 "compare": false, 00:14:12.588 "compare_and_write": false, 00:14:12.588 "abort": true, 00:14:12.588 "seek_hole": false, 00:14:12.588 "seek_data": false, 00:14:12.588 "copy": true, 00:14:12.588 "nvme_iov_md": false 00:14:12.588 }, 00:14:12.588 "memory_domains": [ 00:14:12.588 { 00:14:12.588 "dma_device_id": "system", 00:14:12.588 "dma_device_type": 1 00:14:12.588 }, 00:14:12.588 { 00:14:12.588 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:12.588 "dma_device_type": 2 00:14:12.588 } 00:14:12.588 ], 00:14:12.588 "driver_specific": {} 00:14:12.588 } 00:14:12.588 ] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.588 [2024-09-29 16:52:34.111337] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:12.588 [2024-09-29 16:52:34.111800] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:12.588 [2024-09-29 16:52:34.111867] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:12.588 [2024-09-29 16:52:34.113654] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:12.588 [2024-09-29 16:52:34.113762] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.588 "name": "Existed_Raid", 00:14:12.588 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.588 "strip_size_kb": 64, 00:14:12.588 "state": "configuring", 00:14:12.588 "raid_level": "raid5f", 00:14:12.588 "superblock": false, 00:14:12.588 "num_base_bdevs": 4, 00:14:12.588 "num_base_bdevs_discovered": 3, 00:14:12.588 "num_base_bdevs_operational": 4, 00:14:12.588 "base_bdevs_list": [ 00:14:12.588 { 00:14:12.588 "name": "BaseBdev1", 00:14:12.588 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.588 "is_configured": false, 00:14:12.588 "data_offset": 0, 00:14:12.588 "data_size": 0 00:14:12.588 }, 00:14:12.588 { 00:14:12.588 "name": "BaseBdev2", 00:14:12.588 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:12.588 "is_configured": true, 00:14:12.588 "data_offset": 0, 00:14:12.588 "data_size": 65536 00:14:12.588 }, 00:14:12.588 { 00:14:12.588 "name": "BaseBdev3", 00:14:12.588 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:12.588 "is_configured": true, 00:14:12.588 "data_offset": 0, 00:14:12.588 "data_size": 65536 00:14:12.588 }, 00:14:12.588 { 00:14:12.588 "name": "BaseBdev4", 00:14:12.588 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:12.588 "is_configured": true, 00:14:12.588 "data_offset": 0, 00:14:12.588 "data_size": 65536 00:14:12.588 } 00:14:12.588 ] 00:14:12.588 }' 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.588 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.158 [2024-09-29 16:52:34.570572] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.158 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.159 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:13.159 "name": "Existed_Raid", 00:14:13.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:13.159 "strip_size_kb": 64, 00:14:13.159 "state": "configuring", 00:14:13.159 "raid_level": "raid5f", 00:14:13.159 "superblock": false, 00:14:13.159 "num_base_bdevs": 4, 00:14:13.159 "num_base_bdevs_discovered": 2, 00:14:13.159 "num_base_bdevs_operational": 4, 00:14:13.159 "base_bdevs_list": [ 00:14:13.159 { 00:14:13.159 "name": "BaseBdev1", 00:14:13.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:13.159 "is_configured": false, 00:14:13.159 "data_offset": 0, 00:14:13.159 "data_size": 0 00:14:13.159 }, 00:14:13.159 { 00:14:13.159 "name": null, 00:14:13.159 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:13.159 "is_configured": false, 00:14:13.159 "data_offset": 0, 00:14:13.159 "data_size": 65536 00:14:13.159 }, 00:14:13.159 { 00:14:13.159 "name": "BaseBdev3", 00:14:13.159 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:13.159 "is_configured": true, 00:14:13.159 "data_offset": 0, 00:14:13.159 "data_size": 65536 00:14:13.159 }, 00:14:13.159 { 00:14:13.159 "name": "BaseBdev4", 00:14:13.159 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:13.159 "is_configured": true, 00:14:13.159 "data_offset": 0, 00:14:13.159 "data_size": 65536 00:14:13.159 } 00:14:13.159 ] 00:14:13.159 }' 00:14:13.159 16:52:34 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:13.159 16:52:34 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.419 BaseBdev1 00:14:13.419 [2024-09-29 16:52:35.080815] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.419 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.679 [ 00:14:13.679 { 00:14:13.679 "name": "BaseBdev1", 00:14:13.679 "aliases": [ 00:14:13.679 "4707af93-b152-4455-982d-3a08877bebb1" 00:14:13.679 ], 00:14:13.679 "product_name": "Malloc disk", 00:14:13.679 "block_size": 512, 00:14:13.679 "num_blocks": 65536, 00:14:13.679 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:13.679 "assigned_rate_limits": { 00:14:13.679 "rw_ios_per_sec": 0, 00:14:13.679 "rw_mbytes_per_sec": 0, 00:14:13.679 "r_mbytes_per_sec": 0, 00:14:13.679 "w_mbytes_per_sec": 0 00:14:13.679 }, 00:14:13.679 "claimed": true, 00:14:13.679 "claim_type": "exclusive_write", 00:14:13.679 "zoned": false, 00:14:13.679 "supported_io_types": { 00:14:13.679 "read": true, 00:14:13.679 "write": true, 00:14:13.679 "unmap": true, 00:14:13.679 "flush": true, 00:14:13.679 "reset": true, 00:14:13.679 "nvme_admin": false, 00:14:13.679 "nvme_io": false, 00:14:13.679 "nvme_io_md": false, 00:14:13.679 "write_zeroes": true, 00:14:13.679 "zcopy": true, 00:14:13.679 "get_zone_info": false, 00:14:13.679 "zone_management": false, 00:14:13.679 "zone_append": false, 00:14:13.679 "compare": false, 00:14:13.679 "compare_and_write": false, 00:14:13.679 "abort": true, 00:14:13.679 "seek_hole": false, 00:14:13.679 "seek_data": false, 00:14:13.679 "copy": true, 00:14:13.679 "nvme_iov_md": false 00:14:13.679 }, 00:14:13.679 "memory_domains": [ 00:14:13.679 { 00:14:13.679 "dma_device_id": "system", 00:14:13.679 "dma_device_type": 1 00:14:13.679 }, 00:14:13.679 { 00:14:13.679 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:13.679 "dma_device_type": 2 00:14:13.679 } 00:14:13.679 ], 00:14:13.679 "driver_specific": {} 00:14:13.679 } 00:14:13.679 ] 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:13.679 "name": "Existed_Raid", 00:14:13.679 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:13.679 "strip_size_kb": 64, 00:14:13.679 "state": "configuring", 00:14:13.679 "raid_level": "raid5f", 00:14:13.679 "superblock": false, 00:14:13.679 "num_base_bdevs": 4, 00:14:13.679 "num_base_bdevs_discovered": 3, 00:14:13.679 "num_base_bdevs_operational": 4, 00:14:13.679 "base_bdevs_list": [ 00:14:13.679 { 00:14:13.679 "name": "BaseBdev1", 00:14:13.679 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:13.679 "is_configured": true, 00:14:13.679 "data_offset": 0, 00:14:13.679 "data_size": 65536 00:14:13.679 }, 00:14:13.679 { 00:14:13.679 "name": null, 00:14:13.679 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:13.679 "is_configured": false, 00:14:13.679 "data_offset": 0, 00:14:13.679 "data_size": 65536 00:14:13.679 }, 00:14:13.679 { 00:14:13.679 "name": "BaseBdev3", 00:14:13.679 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:13.679 "is_configured": true, 00:14:13.679 "data_offset": 0, 00:14:13.679 "data_size": 65536 00:14:13.679 }, 00:14:13.679 { 00:14:13.679 "name": "BaseBdev4", 00:14:13.679 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:13.679 "is_configured": true, 00:14:13.679 "data_offset": 0, 00:14:13.679 "data_size": 65536 00:14:13.679 } 00:14:13.679 ] 00:14:13.679 }' 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:13.679 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.939 [2024-09-29 16:52:35.599998] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.939 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.199 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:14.199 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.199 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.199 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.199 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.199 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.199 "name": "Existed_Raid", 00:14:14.199 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.199 "strip_size_kb": 64, 00:14:14.199 "state": "configuring", 00:14:14.199 "raid_level": "raid5f", 00:14:14.199 "superblock": false, 00:14:14.199 "num_base_bdevs": 4, 00:14:14.199 "num_base_bdevs_discovered": 2, 00:14:14.199 "num_base_bdevs_operational": 4, 00:14:14.199 "base_bdevs_list": [ 00:14:14.199 { 00:14:14.199 "name": "BaseBdev1", 00:14:14.199 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:14.199 "is_configured": true, 00:14:14.199 "data_offset": 0, 00:14:14.199 "data_size": 65536 00:14:14.199 }, 00:14:14.199 { 00:14:14.199 "name": null, 00:14:14.199 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:14.199 "is_configured": false, 00:14:14.199 "data_offset": 0, 00:14:14.199 "data_size": 65536 00:14:14.199 }, 00:14:14.199 { 00:14:14.199 "name": null, 00:14:14.199 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:14.199 "is_configured": false, 00:14:14.199 "data_offset": 0, 00:14:14.199 "data_size": 65536 00:14:14.199 }, 00:14:14.199 { 00:14:14.199 "name": "BaseBdev4", 00:14:14.199 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:14.199 "is_configured": true, 00:14:14.199 "data_offset": 0, 00:14:14.200 "data_size": 65536 00:14:14.200 } 00:14:14.200 ] 00:14:14.200 }' 00:14:14.200 16:52:35 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.200 16:52:35 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.459 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.459 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:14.459 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.459 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.460 [2024-09-29 16:52:36.103258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.460 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.720 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.720 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:14.720 "name": "Existed_Raid", 00:14:14.720 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:14.720 "strip_size_kb": 64, 00:14:14.720 "state": "configuring", 00:14:14.720 "raid_level": "raid5f", 00:14:14.720 "superblock": false, 00:14:14.720 "num_base_bdevs": 4, 00:14:14.720 "num_base_bdevs_discovered": 3, 00:14:14.720 "num_base_bdevs_operational": 4, 00:14:14.720 "base_bdevs_list": [ 00:14:14.720 { 00:14:14.720 "name": "BaseBdev1", 00:14:14.720 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:14.720 "is_configured": true, 00:14:14.720 "data_offset": 0, 00:14:14.720 "data_size": 65536 00:14:14.720 }, 00:14:14.720 { 00:14:14.720 "name": null, 00:14:14.720 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:14.720 "is_configured": false, 00:14:14.720 "data_offset": 0, 00:14:14.720 "data_size": 65536 00:14:14.720 }, 00:14:14.720 { 00:14:14.720 "name": "BaseBdev3", 00:14:14.720 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:14.720 "is_configured": true, 00:14:14.720 "data_offset": 0, 00:14:14.720 "data_size": 65536 00:14:14.720 }, 00:14:14.720 { 00:14:14.720 "name": "BaseBdev4", 00:14:14.720 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:14.720 "is_configured": true, 00:14:14.720 "data_offset": 0, 00:14:14.720 "data_size": 65536 00:14:14.720 } 00:14:14.720 ] 00:14:14.720 }' 00:14:14.720 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:14.720 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.980 [2024-09-29 16:52:36.614356] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:14.980 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:14.981 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:14.981 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:14.981 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:14.981 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:14.981 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.981 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.240 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.240 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.240 "name": "Existed_Raid", 00:14:15.240 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.240 "strip_size_kb": 64, 00:14:15.240 "state": "configuring", 00:14:15.240 "raid_level": "raid5f", 00:14:15.240 "superblock": false, 00:14:15.240 "num_base_bdevs": 4, 00:14:15.240 "num_base_bdevs_discovered": 2, 00:14:15.240 "num_base_bdevs_operational": 4, 00:14:15.240 "base_bdevs_list": [ 00:14:15.240 { 00:14:15.240 "name": null, 00:14:15.240 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:15.240 "is_configured": false, 00:14:15.240 "data_offset": 0, 00:14:15.240 "data_size": 65536 00:14:15.240 }, 00:14:15.240 { 00:14:15.240 "name": null, 00:14:15.240 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:15.240 "is_configured": false, 00:14:15.240 "data_offset": 0, 00:14:15.240 "data_size": 65536 00:14:15.240 }, 00:14:15.240 { 00:14:15.240 "name": "BaseBdev3", 00:14:15.240 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:15.240 "is_configured": true, 00:14:15.240 "data_offset": 0, 00:14:15.240 "data_size": 65536 00:14:15.240 }, 00:14:15.240 { 00:14:15.240 "name": "BaseBdev4", 00:14:15.240 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:15.240 "is_configured": true, 00:14:15.240 "data_offset": 0, 00:14:15.240 "data_size": 65536 00:14:15.240 } 00:14:15.240 ] 00:14:15.240 }' 00:14:15.240 16:52:36 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.240 16:52:36 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.501 [2024-09-29 16:52:37.152156] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:15.501 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.761 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.761 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.761 "name": "Existed_Raid", 00:14:15.761 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:15.761 "strip_size_kb": 64, 00:14:15.761 "state": "configuring", 00:14:15.761 "raid_level": "raid5f", 00:14:15.761 "superblock": false, 00:14:15.761 "num_base_bdevs": 4, 00:14:15.761 "num_base_bdevs_discovered": 3, 00:14:15.761 "num_base_bdevs_operational": 4, 00:14:15.761 "base_bdevs_list": [ 00:14:15.761 { 00:14:15.761 "name": null, 00:14:15.761 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:15.761 "is_configured": false, 00:14:15.761 "data_offset": 0, 00:14:15.761 "data_size": 65536 00:14:15.761 }, 00:14:15.761 { 00:14:15.761 "name": "BaseBdev2", 00:14:15.761 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:15.761 "is_configured": true, 00:14:15.761 "data_offset": 0, 00:14:15.761 "data_size": 65536 00:14:15.761 }, 00:14:15.761 { 00:14:15.761 "name": "BaseBdev3", 00:14:15.761 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:15.761 "is_configured": true, 00:14:15.761 "data_offset": 0, 00:14:15.761 "data_size": 65536 00:14:15.761 }, 00:14:15.761 { 00:14:15.761 "name": "BaseBdev4", 00:14:15.761 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:15.761 "is_configured": true, 00:14:15.761 "data_offset": 0, 00:14:15.761 "data_size": 65536 00:14:15.761 } 00:14:15.761 ] 00:14:15.761 }' 00:14:15.761 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.761 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 4707af93-b152-4455-982d-3a08877bebb1 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.021 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.281 NewBaseBdev 00:14:16.281 [2024-09-29 16:52:37.694116] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:16.281 [2024-09-29 16:52:37.694164] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:16.281 [2024-09-29 16:52:37.694171] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:16.281 [2024-09-29 16:52:37.694437] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:16.281 [2024-09-29 16:52:37.694877] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:16.281 [2024-09-29 16:52:37.694892] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:16.281 [2024-09-29 16:52:37.695063] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.281 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.281 [ 00:14:16.281 { 00:14:16.281 "name": "NewBaseBdev", 00:14:16.281 "aliases": [ 00:14:16.281 "4707af93-b152-4455-982d-3a08877bebb1" 00:14:16.281 ], 00:14:16.281 "product_name": "Malloc disk", 00:14:16.281 "block_size": 512, 00:14:16.281 "num_blocks": 65536, 00:14:16.281 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:16.281 "assigned_rate_limits": { 00:14:16.281 "rw_ios_per_sec": 0, 00:14:16.281 "rw_mbytes_per_sec": 0, 00:14:16.281 "r_mbytes_per_sec": 0, 00:14:16.281 "w_mbytes_per_sec": 0 00:14:16.281 }, 00:14:16.281 "claimed": true, 00:14:16.281 "claim_type": "exclusive_write", 00:14:16.281 "zoned": false, 00:14:16.281 "supported_io_types": { 00:14:16.282 "read": true, 00:14:16.282 "write": true, 00:14:16.282 "unmap": true, 00:14:16.282 "flush": true, 00:14:16.282 "reset": true, 00:14:16.282 "nvme_admin": false, 00:14:16.282 "nvme_io": false, 00:14:16.282 "nvme_io_md": false, 00:14:16.282 "write_zeroes": true, 00:14:16.282 "zcopy": true, 00:14:16.282 "get_zone_info": false, 00:14:16.282 "zone_management": false, 00:14:16.282 "zone_append": false, 00:14:16.282 "compare": false, 00:14:16.282 "compare_and_write": false, 00:14:16.282 "abort": true, 00:14:16.282 "seek_hole": false, 00:14:16.282 "seek_data": false, 00:14:16.282 "copy": true, 00:14:16.282 "nvme_iov_md": false 00:14:16.282 }, 00:14:16.282 "memory_domains": [ 00:14:16.282 { 00:14:16.282 "dma_device_id": "system", 00:14:16.282 "dma_device_type": 1 00:14:16.282 }, 00:14:16.282 { 00:14:16.282 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:16.282 "dma_device_type": 2 00:14:16.282 } 00:14:16.282 ], 00:14:16.282 "driver_specific": {} 00:14:16.282 } 00:14:16.282 ] 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.282 "name": "Existed_Raid", 00:14:16.282 "uuid": "d77a469c-9f1c-4e47-ab38-404237c8d94a", 00:14:16.282 "strip_size_kb": 64, 00:14:16.282 "state": "online", 00:14:16.282 "raid_level": "raid5f", 00:14:16.282 "superblock": false, 00:14:16.282 "num_base_bdevs": 4, 00:14:16.282 "num_base_bdevs_discovered": 4, 00:14:16.282 "num_base_bdevs_operational": 4, 00:14:16.282 "base_bdevs_list": [ 00:14:16.282 { 00:14:16.282 "name": "NewBaseBdev", 00:14:16.282 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:16.282 "is_configured": true, 00:14:16.282 "data_offset": 0, 00:14:16.282 "data_size": 65536 00:14:16.282 }, 00:14:16.282 { 00:14:16.282 "name": "BaseBdev2", 00:14:16.282 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:16.282 "is_configured": true, 00:14:16.282 "data_offset": 0, 00:14:16.282 "data_size": 65536 00:14:16.282 }, 00:14:16.282 { 00:14:16.282 "name": "BaseBdev3", 00:14:16.282 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:16.282 "is_configured": true, 00:14:16.282 "data_offset": 0, 00:14:16.282 "data_size": 65536 00:14:16.282 }, 00:14:16.282 { 00:14:16.282 "name": "BaseBdev4", 00:14:16.282 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:16.282 "is_configured": true, 00:14:16.282 "data_offset": 0, 00:14:16.282 "data_size": 65536 00:14:16.282 } 00:14:16.282 ] 00:14:16.282 }' 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.282 16:52:37 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.542 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.542 [2024-09-29 16:52:38.209455] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:16.801 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.801 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:16.801 "name": "Existed_Raid", 00:14:16.801 "aliases": [ 00:14:16.801 "d77a469c-9f1c-4e47-ab38-404237c8d94a" 00:14:16.801 ], 00:14:16.801 "product_name": "Raid Volume", 00:14:16.801 "block_size": 512, 00:14:16.801 "num_blocks": 196608, 00:14:16.801 "uuid": "d77a469c-9f1c-4e47-ab38-404237c8d94a", 00:14:16.801 "assigned_rate_limits": { 00:14:16.801 "rw_ios_per_sec": 0, 00:14:16.801 "rw_mbytes_per_sec": 0, 00:14:16.801 "r_mbytes_per_sec": 0, 00:14:16.801 "w_mbytes_per_sec": 0 00:14:16.801 }, 00:14:16.801 "claimed": false, 00:14:16.801 "zoned": false, 00:14:16.801 "supported_io_types": { 00:14:16.801 "read": true, 00:14:16.801 "write": true, 00:14:16.801 "unmap": false, 00:14:16.801 "flush": false, 00:14:16.801 "reset": true, 00:14:16.801 "nvme_admin": false, 00:14:16.801 "nvme_io": false, 00:14:16.801 "nvme_io_md": false, 00:14:16.801 "write_zeroes": true, 00:14:16.801 "zcopy": false, 00:14:16.801 "get_zone_info": false, 00:14:16.801 "zone_management": false, 00:14:16.801 "zone_append": false, 00:14:16.801 "compare": false, 00:14:16.801 "compare_and_write": false, 00:14:16.801 "abort": false, 00:14:16.801 "seek_hole": false, 00:14:16.801 "seek_data": false, 00:14:16.801 "copy": false, 00:14:16.801 "nvme_iov_md": false 00:14:16.801 }, 00:14:16.801 "driver_specific": { 00:14:16.801 "raid": { 00:14:16.801 "uuid": "d77a469c-9f1c-4e47-ab38-404237c8d94a", 00:14:16.801 "strip_size_kb": 64, 00:14:16.801 "state": "online", 00:14:16.801 "raid_level": "raid5f", 00:14:16.801 "superblock": false, 00:14:16.801 "num_base_bdevs": 4, 00:14:16.801 "num_base_bdevs_discovered": 4, 00:14:16.801 "num_base_bdevs_operational": 4, 00:14:16.801 "base_bdevs_list": [ 00:14:16.801 { 00:14:16.801 "name": "NewBaseBdev", 00:14:16.801 "uuid": "4707af93-b152-4455-982d-3a08877bebb1", 00:14:16.801 "is_configured": true, 00:14:16.801 "data_offset": 0, 00:14:16.801 "data_size": 65536 00:14:16.801 }, 00:14:16.801 { 00:14:16.801 "name": "BaseBdev2", 00:14:16.801 "uuid": "a885c108-dcc3-4c30-be65-20e23f1d7985", 00:14:16.801 "is_configured": true, 00:14:16.801 "data_offset": 0, 00:14:16.801 "data_size": 65536 00:14:16.801 }, 00:14:16.801 { 00:14:16.801 "name": "BaseBdev3", 00:14:16.801 "uuid": "e40e952b-02ed-4bd4-be62-ff258a75f0b1", 00:14:16.801 "is_configured": true, 00:14:16.801 "data_offset": 0, 00:14:16.801 "data_size": 65536 00:14:16.801 }, 00:14:16.801 { 00:14:16.801 "name": "BaseBdev4", 00:14:16.801 "uuid": "86b75c21-daba-4724-87a7-8f1a82e7098a", 00:14:16.801 "is_configured": true, 00:14:16.801 "data_offset": 0, 00:14:16.801 "data_size": 65536 00:14:16.801 } 00:14:16.801 ] 00:14:16.801 } 00:14:16.801 } 00:14:16.801 }' 00:14:16.801 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:16.801 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:16.802 BaseBdev2 00:14:16.802 BaseBdev3 00:14:16.802 BaseBdev4' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.802 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.062 [2024-09-29 16:52:38.548712] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:17.062 [2024-09-29 16:52:38.548748] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:17.062 [2024-09-29 16:52:38.548815] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:17.062 [2024-09-29 16:52:38.549050] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:17.062 [2024-09-29 16:52:38.549060] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 92894 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 92894 ']' 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 92894 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 92894 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:17.062 killing process with pid 92894 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 92894' 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 92894 00:14:17.062 [2024-09-29 16:52:38.597453] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:17.062 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 92894 00:14:17.062 [2024-09-29 16:52:38.637864] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:17.323 ************************************ 00:14:17.323 END TEST raid5f_state_function_test 00:14:17.323 ************************************ 00:14:17.323 00:14:17.323 real 0m9.919s 00:14:17.323 user 0m16.922s 00:14:17.323 sys 0m2.212s 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:17.323 16:52:38 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:14:17.323 16:52:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:14:17.323 16:52:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:17.323 16:52:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:17.323 ************************************ 00:14:17.323 START TEST raid5f_state_function_test_sb 00:14:17.323 ************************************ 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 true 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:17.323 Process raid pid: 93543 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=93543 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93543' 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 93543 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 93543 ']' 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:17.323 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:17.323 16:52:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:17.583 [2024-09-29 16:52:39.067475] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:14:17.583 [2024-09-29 16:52:39.067659] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:17.583 [2024-09-29 16:52:39.213100] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:17.844 [2024-09-29 16:52:39.259019] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:17.844 [2024-09-29 16:52:39.301885] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:17.844 [2024-09-29 16:52:39.301918] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.414 [2024-09-29 16:52:39.903307] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:18.414 [2024-09-29 16:52:39.903355] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:18.414 [2024-09-29 16:52:39.903366] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:18.414 [2024-09-29 16:52:39.903377] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:18.414 [2024-09-29 16:52:39.903383] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:18.414 [2024-09-29 16:52:39.903393] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:18.414 [2024-09-29 16:52:39.903399] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:18.414 [2024-09-29 16:52:39.903407] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:18.414 "name": "Existed_Raid", 00:14:18.414 "uuid": "146ba0d7-5698-478c-8502-c7ccbc5fa4fe", 00:14:18.414 "strip_size_kb": 64, 00:14:18.414 "state": "configuring", 00:14:18.414 "raid_level": "raid5f", 00:14:18.414 "superblock": true, 00:14:18.414 "num_base_bdevs": 4, 00:14:18.414 "num_base_bdevs_discovered": 0, 00:14:18.414 "num_base_bdevs_operational": 4, 00:14:18.414 "base_bdevs_list": [ 00:14:18.414 { 00:14:18.414 "name": "BaseBdev1", 00:14:18.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.414 "is_configured": false, 00:14:18.414 "data_offset": 0, 00:14:18.414 "data_size": 0 00:14:18.414 }, 00:14:18.414 { 00:14:18.414 "name": "BaseBdev2", 00:14:18.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.414 "is_configured": false, 00:14:18.414 "data_offset": 0, 00:14:18.414 "data_size": 0 00:14:18.414 }, 00:14:18.414 { 00:14:18.414 "name": "BaseBdev3", 00:14:18.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.414 "is_configured": false, 00:14:18.414 "data_offset": 0, 00:14:18.414 "data_size": 0 00:14:18.414 }, 00:14:18.414 { 00:14:18.414 "name": "BaseBdev4", 00:14:18.414 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.414 "is_configured": false, 00:14:18.414 "data_offset": 0, 00:14:18.414 "data_size": 0 00:14:18.414 } 00:14:18.414 ] 00:14:18.414 }' 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:18.414 16:52:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.986 [2024-09-29 16:52:40.398286] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:18.986 [2024-09-29 16:52:40.398368] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.986 [2024-09-29 16:52:40.410293] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:18.986 [2024-09-29 16:52:40.410384] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:18.986 [2024-09-29 16:52:40.410409] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:18.986 [2024-09-29 16:52:40.410432] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:18.986 [2024-09-29 16:52:40.410450] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:18.986 [2024-09-29 16:52:40.410471] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:18.986 [2024-09-29 16:52:40.410488] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:18.986 [2024-09-29 16:52:40.410531] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.986 [2024-09-29 16:52:40.431308] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:18.986 BaseBdev1 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.986 [ 00:14:18.986 { 00:14:18.986 "name": "BaseBdev1", 00:14:18.986 "aliases": [ 00:14:18.986 "84dd9b33-89a0-479d-a2fb-27809f92eeee" 00:14:18.986 ], 00:14:18.986 "product_name": "Malloc disk", 00:14:18.986 "block_size": 512, 00:14:18.986 "num_blocks": 65536, 00:14:18.986 "uuid": "84dd9b33-89a0-479d-a2fb-27809f92eeee", 00:14:18.986 "assigned_rate_limits": { 00:14:18.986 "rw_ios_per_sec": 0, 00:14:18.986 "rw_mbytes_per_sec": 0, 00:14:18.986 "r_mbytes_per_sec": 0, 00:14:18.986 "w_mbytes_per_sec": 0 00:14:18.986 }, 00:14:18.986 "claimed": true, 00:14:18.986 "claim_type": "exclusive_write", 00:14:18.986 "zoned": false, 00:14:18.986 "supported_io_types": { 00:14:18.986 "read": true, 00:14:18.986 "write": true, 00:14:18.986 "unmap": true, 00:14:18.986 "flush": true, 00:14:18.986 "reset": true, 00:14:18.986 "nvme_admin": false, 00:14:18.986 "nvme_io": false, 00:14:18.986 "nvme_io_md": false, 00:14:18.986 "write_zeroes": true, 00:14:18.986 "zcopy": true, 00:14:18.986 "get_zone_info": false, 00:14:18.986 "zone_management": false, 00:14:18.986 "zone_append": false, 00:14:18.986 "compare": false, 00:14:18.986 "compare_and_write": false, 00:14:18.986 "abort": true, 00:14:18.986 "seek_hole": false, 00:14:18.986 "seek_data": false, 00:14:18.986 "copy": true, 00:14:18.986 "nvme_iov_md": false 00:14:18.986 }, 00:14:18.986 "memory_domains": [ 00:14:18.986 { 00:14:18.986 "dma_device_id": "system", 00:14:18.986 "dma_device_type": 1 00:14:18.986 }, 00:14:18.986 { 00:14:18.986 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:18.986 "dma_device_type": 2 00:14:18.986 } 00:14:18.986 ], 00:14:18.986 "driver_specific": {} 00:14:18.986 } 00:14:18.986 ] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.986 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:18.986 "name": "Existed_Raid", 00:14:18.986 "uuid": "7f2e11bd-4bbc-42cf-b0eb-d5202f384375", 00:14:18.986 "strip_size_kb": 64, 00:14:18.986 "state": "configuring", 00:14:18.986 "raid_level": "raid5f", 00:14:18.986 "superblock": true, 00:14:18.986 "num_base_bdevs": 4, 00:14:18.986 "num_base_bdevs_discovered": 1, 00:14:18.986 "num_base_bdevs_operational": 4, 00:14:18.987 "base_bdevs_list": [ 00:14:18.987 { 00:14:18.987 "name": "BaseBdev1", 00:14:18.987 "uuid": "84dd9b33-89a0-479d-a2fb-27809f92eeee", 00:14:18.987 "is_configured": true, 00:14:18.987 "data_offset": 2048, 00:14:18.987 "data_size": 63488 00:14:18.987 }, 00:14:18.987 { 00:14:18.987 "name": "BaseBdev2", 00:14:18.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.987 "is_configured": false, 00:14:18.987 "data_offset": 0, 00:14:18.987 "data_size": 0 00:14:18.987 }, 00:14:18.987 { 00:14:18.987 "name": "BaseBdev3", 00:14:18.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.987 "is_configured": false, 00:14:18.987 "data_offset": 0, 00:14:18.987 "data_size": 0 00:14:18.987 }, 00:14:18.987 { 00:14:18.987 "name": "BaseBdev4", 00:14:18.987 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.987 "is_configured": false, 00:14:18.987 "data_offset": 0, 00:14:18.987 "data_size": 0 00:14:18.987 } 00:14:18.987 ] 00:14:18.987 }' 00:14:18.987 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:18.987 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.558 [2024-09-29 16:52:40.930496] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:19.558 [2024-09-29 16:52:40.930573] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.558 [2024-09-29 16:52:40.938557] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:19.558 [2024-09-29 16:52:40.940433] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:19.558 [2024-09-29 16:52:40.940505] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:19.558 [2024-09-29 16:52:40.940532] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:19.558 [2024-09-29 16:52:40.940553] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:19.558 [2024-09-29 16:52:40.940570] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:19.558 [2024-09-29 16:52:40.940589] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:19.558 "name": "Existed_Raid", 00:14:19.558 "uuid": "74cf4b91-83f0-4acb-875b-c00552e5cf6f", 00:14:19.558 "strip_size_kb": 64, 00:14:19.558 "state": "configuring", 00:14:19.558 "raid_level": "raid5f", 00:14:19.558 "superblock": true, 00:14:19.558 "num_base_bdevs": 4, 00:14:19.558 "num_base_bdevs_discovered": 1, 00:14:19.558 "num_base_bdevs_operational": 4, 00:14:19.558 "base_bdevs_list": [ 00:14:19.558 { 00:14:19.558 "name": "BaseBdev1", 00:14:19.558 "uuid": "84dd9b33-89a0-479d-a2fb-27809f92eeee", 00:14:19.558 "is_configured": true, 00:14:19.558 "data_offset": 2048, 00:14:19.558 "data_size": 63488 00:14:19.558 }, 00:14:19.558 { 00:14:19.558 "name": "BaseBdev2", 00:14:19.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.558 "is_configured": false, 00:14:19.558 "data_offset": 0, 00:14:19.558 "data_size": 0 00:14:19.558 }, 00:14:19.558 { 00:14:19.558 "name": "BaseBdev3", 00:14:19.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.558 "is_configured": false, 00:14:19.558 "data_offset": 0, 00:14:19.558 "data_size": 0 00:14:19.558 }, 00:14:19.558 { 00:14:19.558 "name": "BaseBdev4", 00:14:19.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:19.558 "is_configured": false, 00:14:19.558 "data_offset": 0, 00:14:19.558 "data_size": 0 00:14:19.558 } 00:14:19.558 ] 00:14:19.558 }' 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:19.558 16:52:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.818 [2024-09-29 16:52:41.460361] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:19.818 BaseBdev2 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.818 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:19.819 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.819 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:19.819 [ 00:14:19.819 { 00:14:19.819 "name": "BaseBdev2", 00:14:19.819 "aliases": [ 00:14:20.079 "72ec027e-51c3-4c96-96b3-c6952ec8d1f8" 00:14:20.079 ], 00:14:20.079 "product_name": "Malloc disk", 00:14:20.079 "block_size": 512, 00:14:20.079 "num_blocks": 65536, 00:14:20.079 "uuid": "72ec027e-51c3-4c96-96b3-c6952ec8d1f8", 00:14:20.079 "assigned_rate_limits": { 00:14:20.079 "rw_ios_per_sec": 0, 00:14:20.079 "rw_mbytes_per_sec": 0, 00:14:20.079 "r_mbytes_per_sec": 0, 00:14:20.079 "w_mbytes_per_sec": 0 00:14:20.079 }, 00:14:20.079 "claimed": true, 00:14:20.079 "claim_type": "exclusive_write", 00:14:20.079 "zoned": false, 00:14:20.079 "supported_io_types": { 00:14:20.079 "read": true, 00:14:20.079 "write": true, 00:14:20.079 "unmap": true, 00:14:20.079 "flush": true, 00:14:20.079 "reset": true, 00:14:20.079 "nvme_admin": false, 00:14:20.079 "nvme_io": false, 00:14:20.079 "nvme_io_md": false, 00:14:20.079 "write_zeroes": true, 00:14:20.079 "zcopy": true, 00:14:20.079 "get_zone_info": false, 00:14:20.079 "zone_management": false, 00:14:20.079 "zone_append": false, 00:14:20.079 "compare": false, 00:14:20.079 "compare_and_write": false, 00:14:20.079 "abort": true, 00:14:20.079 "seek_hole": false, 00:14:20.079 "seek_data": false, 00:14:20.079 "copy": true, 00:14:20.079 "nvme_iov_md": false 00:14:20.079 }, 00:14:20.079 "memory_domains": [ 00:14:20.079 { 00:14:20.079 "dma_device_id": "system", 00:14:20.079 "dma_device_type": 1 00:14:20.079 }, 00:14:20.079 { 00:14:20.079 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:20.079 "dma_device_type": 2 00:14:20.079 } 00:14:20.079 ], 00:14:20.079 "driver_specific": {} 00:14:20.079 } 00:14:20.079 ] 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.079 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:20.079 "name": "Existed_Raid", 00:14:20.079 "uuid": "74cf4b91-83f0-4acb-875b-c00552e5cf6f", 00:14:20.079 "strip_size_kb": 64, 00:14:20.079 "state": "configuring", 00:14:20.079 "raid_level": "raid5f", 00:14:20.079 "superblock": true, 00:14:20.079 "num_base_bdevs": 4, 00:14:20.079 "num_base_bdevs_discovered": 2, 00:14:20.079 "num_base_bdevs_operational": 4, 00:14:20.079 "base_bdevs_list": [ 00:14:20.079 { 00:14:20.079 "name": "BaseBdev1", 00:14:20.079 "uuid": "84dd9b33-89a0-479d-a2fb-27809f92eeee", 00:14:20.079 "is_configured": true, 00:14:20.079 "data_offset": 2048, 00:14:20.079 "data_size": 63488 00:14:20.079 }, 00:14:20.079 { 00:14:20.079 "name": "BaseBdev2", 00:14:20.079 "uuid": "72ec027e-51c3-4c96-96b3-c6952ec8d1f8", 00:14:20.079 "is_configured": true, 00:14:20.079 "data_offset": 2048, 00:14:20.079 "data_size": 63488 00:14:20.079 }, 00:14:20.079 { 00:14:20.079 "name": "BaseBdev3", 00:14:20.079 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.079 "is_configured": false, 00:14:20.080 "data_offset": 0, 00:14:20.080 "data_size": 0 00:14:20.080 }, 00:14:20.080 { 00:14:20.080 "name": "BaseBdev4", 00:14:20.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.080 "is_configured": false, 00:14:20.080 "data_offset": 0, 00:14:20.080 "data_size": 0 00:14:20.080 } 00:14:20.080 ] 00:14:20.080 }' 00:14:20.080 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:20.080 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.340 [2024-09-29 16:52:41.970528] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:20.340 BaseBdev3 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.340 16:52:41 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.340 [ 00:14:20.340 { 00:14:20.340 "name": "BaseBdev3", 00:14:20.340 "aliases": [ 00:14:20.340 "9729409e-0c15-42f4-a2f2-54213974d74e" 00:14:20.340 ], 00:14:20.340 "product_name": "Malloc disk", 00:14:20.340 "block_size": 512, 00:14:20.340 "num_blocks": 65536, 00:14:20.340 "uuid": "9729409e-0c15-42f4-a2f2-54213974d74e", 00:14:20.340 "assigned_rate_limits": { 00:14:20.340 "rw_ios_per_sec": 0, 00:14:20.340 "rw_mbytes_per_sec": 0, 00:14:20.340 "r_mbytes_per_sec": 0, 00:14:20.340 "w_mbytes_per_sec": 0 00:14:20.340 }, 00:14:20.340 "claimed": true, 00:14:20.340 "claim_type": "exclusive_write", 00:14:20.340 "zoned": false, 00:14:20.340 "supported_io_types": { 00:14:20.340 "read": true, 00:14:20.340 "write": true, 00:14:20.340 "unmap": true, 00:14:20.340 "flush": true, 00:14:20.340 "reset": true, 00:14:20.340 "nvme_admin": false, 00:14:20.340 "nvme_io": false, 00:14:20.340 "nvme_io_md": false, 00:14:20.340 "write_zeroes": true, 00:14:20.340 "zcopy": true, 00:14:20.340 "get_zone_info": false, 00:14:20.340 "zone_management": false, 00:14:20.340 "zone_append": false, 00:14:20.340 "compare": false, 00:14:20.340 "compare_and_write": false, 00:14:20.340 "abort": true, 00:14:20.340 "seek_hole": false, 00:14:20.341 "seek_data": false, 00:14:20.341 "copy": true, 00:14:20.341 "nvme_iov_md": false 00:14:20.341 }, 00:14:20.341 "memory_domains": [ 00:14:20.341 { 00:14:20.341 "dma_device_id": "system", 00:14:20.341 "dma_device_type": 1 00:14:20.341 }, 00:14:20.341 { 00:14:20.341 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:20.341 "dma_device_type": 2 00:14:20.341 } 00:14:20.341 ], 00:14:20.341 "driver_specific": {} 00:14:20.341 } 00:14:20.341 ] 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.341 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:20.601 "name": "Existed_Raid", 00:14:20.601 "uuid": "74cf4b91-83f0-4acb-875b-c00552e5cf6f", 00:14:20.601 "strip_size_kb": 64, 00:14:20.601 "state": "configuring", 00:14:20.601 "raid_level": "raid5f", 00:14:20.601 "superblock": true, 00:14:20.601 "num_base_bdevs": 4, 00:14:20.601 "num_base_bdevs_discovered": 3, 00:14:20.601 "num_base_bdevs_operational": 4, 00:14:20.601 "base_bdevs_list": [ 00:14:20.601 { 00:14:20.601 "name": "BaseBdev1", 00:14:20.601 "uuid": "84dd9b33-89a0-479d-a2fb-27809f92eeee", 00:14:20.601 "is_configured": true, 00:14:20.601 "data_offset": 2048, 00:14:20.601 "data_size": 63488 00:14:20.601 }, 00:14:20.601 { 00:14:20.601 "name": "BaseBdev2", 00:14:20.601 "uuid": "72ec027e-51c3-4c96-96b3-c6952ec8d1f8", 00:14:20.601 "is_configured": true, 00:14:20.601 "data_offset": 2048, 00:14:20.601 "data_size": 63488 00:14:20.601 }, 00:14:20.601 { 00:14:20.601 "name": "BaseBdev3", 00:14:20.601 "uuid": "9729409e-0c15-42f4-a2f2-54213974d74e", 00:14:20.601 "is_configured": true, 00:14:20.601 "data_offset": 2048, 00:14:20.601 "data_size": 63488 00:14:20.601 }, 00:14:20.601 { 00:14:20.601 "name": "BaseBdev4", 00:14:20.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:20.601 "is_configured": false, 00:14:20.601 "data_offset": 0, 00:14:20.601 "data_size": 0 00:14:20.601 } 00:14:20.601 ] 00:14:20.601 }' 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:20.601 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.861 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.862 [2024-09-29 16:52:42.480861] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:20.862 [2024-09-29 16:52:42.481079] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:20.862 [2024-09-29 16:52:42.481094] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:20.862 BaseBdev4 00:14:20.862 [2024-09-29 16:52:42.481387] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:20.862 [2024-09-29 16:52:42.481854] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:20.862 [2024-09-29 16:52:42.481875] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:20.862 [2024-09-29 16:52:42.482002] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:20.862 [ 00:14:20.862 { 00:14:20.862 "name": "BaseBdev4", 00:14:20.862 "aliases": [ 00:14:20.862 "b0e18ced-c5b9-4d85-9b17-5214b1a46f61" 00:14:20.862 ], 00:14:20.862 "product_name": "Malloc disk", 00:14:20.862 "block_size": 512, 00:14:20.862 "num_blocks": 65536, 00:14:20.862 "uuid": "b0e18ced-c5b9-4d85-9b17-5214b1a46f61", 00:14:20.862 "assigned_rate_limits": { 00:14:20.862 "rw_ios_per_sec": 0, 00:14:20.862 "rw_mbytes_per_sec": 0, 00:14:20.862 "r_mbytes_per_sec": 0, 00:14:20.862 "w_mbytes_per_sec": 0 00:14:20.862 }, 00:14:20.862 "claimed": true, 00:14:20.862 "claim_type": "exclusive_write", 00:14:20.862 "zoned": false, 00:14:20.862 "supported_io_types": { 00:14:20.862 "read": true, 00:14:20.862 "write": true, 00:14:20.862 "unmap": true, 00:14:20.862 "flush": true, 00:14:20.862 "reset": true, 00:14:20.862 "nvme_admin": false, 00:14:20.862 "nvme_io": false, 00:14:20.862 "nvme_io_md": false, 00:14:20.862 "write_zeroes": true, 00:14:20.862 "zcopy": true, 00:14:20.862 "get_zone_info": false, 00:14:20.862 "zone_management": false, 00:14:20.862 "zone_append": false, 00:14:20.862 "compare": false, 00:14:20.862 "compare_and_write": false, 00:14:20.862 "abort": true, 00:14:20.862 "seek_hole": false, 00:14:20.862 "seek_data": false, 00:14:20.862 "copy": true, 00:14:20.862 "nvme_iov_md": false 00:14:20.862 }, 00:14:20.862 "memory_domains": [ 00:14:20.862 { 00:14:20.862 "dma_device_id": "system", 00:14:20.862 "dma_device_type": 1 00:14:20.862 }, 00:14:20.862 { 00:14:20.862 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:20.862 "dma_device_type": 2 00:14:20.862 } 00:14:20.862 ], 00:14:20.862 "driver_specific": {} 00:14:20.862 } 00:14:20.862 ] 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:20.862 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.122 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.122 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.122 "name": "Existed_Raid", 00:14:21.122 "uuid": "74cf4b91-83f0-4acb-875b-c00552e5cf6f", 00:14:21.122 "strip_size_kb": 64, 00:14:21.122 "state": "online", 00:14:21.122 "raid_level": "raid5f", 00:14:21.122 "superblock": true, 00:14:21.122 "num_base_bdevs": 4, 00:14:21.122 "num_base_bdevs_discovered": 4, 00:14:21.122 "num_base_bdevs_operational": 4, 00:14:21.122 "base_bdevs_list": [ 00:14:21.122 { 00:14:21.122 "name": "BaseBdev1", 00:14:21.122 "uuid": "84dd9b33-89a0-479d-a2fb-27809f92eeee", 00:14:21.122 "is_configured": true, 00:14:21.122 "data_offset": 2048, 00:14:21.122 "data_size": 63488 00:14:21.122 }, 00:14:21.122 { 00:14:21.122 "name": "BaseBdev2", 00:14:21.122 "uuid": "72ec027e-51c3-4c96-96b3-c6952ec8d1f8", 00:14:21.122 "is_configured": true, 00:14:21.122 "data_offset": 2048, 00:14:21.122 "data_size": 63488 00:14:21.122 }, 00:14:21.122 { 00:14:21.122 "name": "BaseBdev3", 00:14:21.122 "uuid": "9729409e-0c15-42f4-a2f2-54213974d74e", 00:14:21.122 "is_configured": true, 00:14:21.122 "data_offset": 2048, 00:14:21.122 "data_size": 63488 00:14:21.122 }, 00:14:21.122 { 00:14:21.122 "name": "BaseBdev4", 00:14:21.122 "uuid": "b0e18ced-c5b9-4d85-9b17-5214b1a46f61", 00:14:21.122 "is_configured": true, 00:14:21.122 "data_offset": 2048, 00:14:21.122 "data_size": 63488 00:14:21.122 } 00:14:21.122 ] 00:14:21.122 }' 00:14:21.122 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.122 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.382 [2024-09-29 16:52:42.952332] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.382 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:21.382 "name": "Existed_Raid", 00:14:21.382 "aliases": [ 00:14:21.382 "74cf4b91-83f0-4acb-875b-c00552e5cf6f" 00:14:21.382 ], 00:14:21.382 "product_name": "Raid Volume", 00:14:21.382 "block_size": 512, 00:14:21.382 "num_blocks": 190464, 00:14:21.382 "uuid": "74cf4b91-83f0-4acb-875b-c00552e5cf6f", 00:14:21.382 "assigned_rate_limits": { 00:14:21.382 "rw_ios_per_sec": 0, 00:14:21.382 "rw_mbytes_per_sec": 0, 00:14:21.382 "r_mbytes_per_sec": 0, 00:14:21.382 "w_mbytes_per_sec": 0 00:14:21.382 }, 00:14:21.382 "claimed": false, 00:14:21.382 "zoned": false, 00:14:21.382 "supported_io_types": { 00:14:21.382 "read": true, 00:14:21.382 "write": true, 00:14:21.382 "unmap": false, 00:14:21.382 "flush": false, 00:14:21.382 "reset": true, 00:14:21.382 "nvme_admin": false, 00:14:21.382 "nvme_io": false, 00:14:21.382 "nvme_io_md": false, 00:14:21.382 "write_zeroes": true, 00:14:21.382 "zcopy": false, 00:14:21.382 "get_zone_info": false, 00:14:21.382 "zone_management": false, 00:14:21.382 "zone_append": false, 00:14:21.382 "compare": false, 00:14:21.382 "compare_and_write": false, 00:14:21.382 "abort": false, 00:14:21.382 "seek_hole": false, 00:14:21.382 "seek_data": false, 00:14:21.382 "copy": false, 00:14:21.382 "nvme_iov_md": false 00:14:21.382 }, 00:14:21.382 "driver_specific": { 00:14:21.382 "raid": { 00:14:21.382 "uuid": "74cf4b91-83f0-4acb-875b-c00552e5cf6f", 00:14:21.382 "strip_size_kb": 64, 00:14:21.383 "state": "online", 00:14:21.383 "raid_level": "raid5f", 00:14:21.383 "superblock": true, 00:14:21.383 "num_base_bdevs": 4, 00:14:21.383 "num_base_bdevs_discovered": 4, 00:14:21.383 "num_base_bdevs_operational": 4, 00:14:21.383 "base_bdevs_list": [ 00:14:21.383 { 00:14:21.383 "name": "BaseBdev1", 00:14:21.383 "uuid": "84dd9b33-89a0-479d-a2fb-27809f92eeee", 00:14:21.383 "is_configured": true, 00:14:21.383 "data_offset": 2048, 00:14:21.383 "data_size": 63488 00:14:21.383 }, 00:14:21.383 { 00:14:21.383 "name": "BaseBdev2", 00:14:21.383 "uuid": "72ec027e-51c3-4c96-96b3-c6952ec8d1f8", 00:14:21.383 "is_configured": true, 00:14:21.383 "data_offset": 2048, 00:14:21.383 "data_size": 63488 00:14:21.383 }, 00:14:21.383 { 00:14:21.383 "name": "BaseBdev3", 00:14:21.383 "uuid": "9729409e-0c15-42f4-a2f2-54213974d74e", 00:14:21.383 "is_configured": true, 00:14:21.383 "data_offset": 2048, 00:14:21.383 "data_size": 63488 00:14:21.383 }, 00:14:21.383 { 00:14:21.383 "name": "BaseBdev4", 00:14:21.383 "uuid": "b0e18ced-c5b9-4d85-9b17-5214b1a46f61", 00:14:21.383 "is_configured": true, 00:14:21.383 "data_offset": 2048, 00:14:21.383 "data_size": 63488 00:14:21.383 } 00:14:21.383 ] 00:14:21.383 } 00:14:21.383 } 00:14:21.383 }' 00:14:21.383 16:52:42 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:21.383 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:21.383 BaseBdev2 00:14:21.383 BaseBdev3 00:14:21.383 BaseBdev4' 00:14:21.383 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.720 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.721 [2024-09-29 16:52:43.259675] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:21.721 "name": "Existed_Raid", 00:14:21.721 "uuid": "74cf4b91-83f0-4acb-875b-c00552e5cf6f", 00:14:21.721 "strip_size_kb": 64, 00:14:21.721 "state": "online", 00:14:21.721 "raid_level": "raid5f", 00:14:21.721 "superblock": true, 00:14:21.721 "num_base_bdevs": 4, 00:14:21.721 "num_base_bdevs_discovered": 3, 00:14:21.721 "num_base_bdevs_operational": 3, 00:14:21.721 "base_bdevs_list": [ 00:14:21.721 { 00:14:21.721 "name": null, 00:14:21.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:21.721 "is_configured": false, 00:14:21.721 "data_offset": 0, 00:14:21.721 "data_size": 63488 00:14:21.721 }, 00:14:21.721 { 00:14:21.721 "name": "BaseBdev2", 00:14:21.721 "uuid": "72ec027e-51c3-4c96-96b3-c6952ec8d1f8", 00:14:21.721 "is_configured": true, 00:14:21.721 "data_offset": 2048, 00:14:21.721 "data_size": 63488 00:14:21.721 }, 00:14:21.721 { 00:14:21.721 "name": "BaseBdev3", 00:14:21.721 "uuid": "9729409e-0c15-42f4-a2f2-54213974d74e", 00:14:21.721 "is_configured": true, 00:14:21.721 "data_offset": 2048, 00:14:21.721 "data_size": 63488 00:14:21.721 }, 00:14:21.721 { 00:14:21.721 "name": "BaseBdev4", 00:14:21.721 "uuid": "b0e18ced-c5b9-4d85-9b17-5214b1a46f61", 00:14:21.721 "is_configured": true, 00:14:21.721 "data_offset": 2048, 00:14:21.721 "data_size": 63488 00:14:21.721 } 00:14:21.721 ] 00:14:21.721 }' 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:21.721 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 [2024-09-29 16:52:43.754274] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:22.292 [2024-09-29 16:52:43.754412] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:22.292 [2024-09-29 16:52:43.765732] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 [2024-09-29 16:52:43.821658] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 [2024-09-29 16:52:43.892756] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:22.292 [2024-09-29 16:52:43.892847] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:22.292 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:22.293 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.293 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.553 BaseBdev2 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.553 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:22.554 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.554 16:52:43 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.554 [ 00:14:22.554 { 00:14:22.554 "name": "BaseBdev2", 00:14:22.554 "aliases": [ 00:14:22.554 "ebca5f27-a2ae-4cb8-a831-dc724ffc2947" 00:14:22.554 ], 00:14:22.554 "product_name": "Malloc disk", 00:14:22.554 "block_size": 512, 00:14:22.554 "num_blocks": 65536, 00:14:22.554 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:22.554 "assigned_rate_limits": { 00:14:22.554 "rw_ios_per_sec": 0, 00:14:22.554 "rw_mbytes_per_sec": 0, 00:14:22.554 "r_mbytes_per_sec": 0, 00:14:22.554 "w_mbytes_per_sec": 0 00:14:22.554 }, 00:14:22.554 "claimed": false, 00:14:22.554 "zoned": false, 00:14:22.554 "supported_io_types": { 00:14:22.554 "read": true, 00:14:22.554 "write": true, 00:14:22.554 "unmap": true, 00:14:22.554 "flush": true, 00:14:22.554 "reset": true, 00:14:22.554 "nvme_admin": false, 00:14:22.554 "nvme_io": false, 00:14:22.554 "nvme_io_md": false, 00:14:22.554 "write_zeroes": true, 00:14:22.554 "zcopy": true, 00:14:22.554 "get_zone_info": false, 00:14:22.554 "zone_management": false, 00:14:22.554 "zone_append": false, 00:14:22.554 "compare": false, 00:14:22.554 "compare_and_write": false, 00:14:22.554 "abort": true, 00:14:22.554 "seek_hole": false, 00:14:22.554 "seek_data": false, 00:14:22.554 "copy": true, 00:14:22.554 "nvme_iov_md": false 00:14:22.554 }, 00:14:22.554 "memory_domains": [ 00:14:22.554 { 00:14:22.554 "dma_device_id": "system", 00:14:22.554 "dma_device_type": 1 00:14:22.554 }, 00:14:22.554 { 00:14:22.554 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:22.554 "dma_device_type": 2 00:14:22.554 } 00:14:22.554 ], 00:14:22.554 "driver_specific": {} 00:14:22.554 } 00:14:22.554 ] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.554 BaseBdev3 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.554 [ 00:14:22.554 { 00:14:22.554 "name": "BaseBdev3", 00:14:22.554 "aliases": [ 00:14:22.554 "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5" 00:14:22.554 ], 00:14:22.554 "product_name": "Malloc disk", 00:14:22.554 "block_size": 512, 00:14:22.554 "num_blocks": 65536, 00:14:22.554 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:22.554 "assigned_rate_limits": { 00:14:22.554 "rw_ios_per_sec": 0, 00:14:22.554 "rw_mbytes_per_sec": 0, 00:14:22.554 "r_mbytes_per_sec": 0, 00:14:22.554 "w_mbytes_per_sec": 0 00:14:22.554 }, 00:14:22.554 "claimed": false, 00:14:22.554 "zoned": false, 00:14:22.554 "supported_io_types": { 00:14:22.554 "read": true, 00:14:22.554 "write": true, 00:14:22.554 "unmap": true, 00:14:22.554 "flush": true, 00:14:22.554 "reset": true, 00:14:22.554 "nvme_admin": false, 00:14:22.554 "nvme_io": false, 00:14:22.554 "nvme_io_md": false, 00:14:22.554 "write_zeroes": true, 00:14:22.554 "zcopy": true, 00:14:22.554 "get_zone_info": false, 00:14:22.554 "zone_management": false, 00:14:22.554 "zone_append": false, 00:14:22.554 "compare": false, 00:14:22.554 "compare_and_write": false, 00:14:22.554 "abort": true, 00:14:22.554 "seek_hole": false, 00:14:22.554 "seek_data": false, 00:14:22.554 "copy": true, 00:14:22.554 "nvme_iov_md": false 00:14:22.554 }, 00:14:22.554 "memory_domains": [ 00:14:22.554 { 00:14:22.554 "dma_device_id": "system", 00:14:22.554 "dma_device_type": 1 00:14:22.554 }, 00:14:22.554 { 00:14:22.554 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:22.554 "dma_device_type": 2 00:14:22.554 } 00:14:22.554 ], 00:14:22.554 "driver_specific": {} 00:14:22.554 } 00:14:22.554 ] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.554 BaseBdev4 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.554 [ 00:14:22.554 { 00:14:22.554 "name": "BaseBdev4", 00:14:22.554 "aliases": [ 00:14:22.554 "9ea5dd48-9810-425a-ac11-f4cc0cc80999" 00:14:22.554 ], 00:14:22.554 "product_name": "Malloc disk", 00:14:22.554 "block_size": 512, 00:14:22.554 "num_blocks": 65536, 00:14:22.554 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:22.554 "assigned_rate_limits": { 00:14:22.554 "rw_ios_per_sec": 0, 00:14:22.554 "rw_mbytes_per_sec": 0, 00:14:22.554 "r_mbytes_per_sec": 0, 00:14:22.554 "w_mbytes_per_sec": 0 00:14:22.554 }, 00:14:22.554 "claimed": false, 00:14:22.554 "zoned": false, 00:14:22.554 "supported_io_types": { 00:14:22.554 "read": true, 00:14:22.554 "write": true, 00:14:22.554 "unmap": true, 00:14:22.554 "flush": true, 00:14:22.554 "reset": true, 00:14:22.554 "nvme_admin": false, 00:14:22.554 "nvme_io": false, 00:14:22.554 "nvme_io_md": false, 00:14:22.554 "write_zeroes": true, 00:14:22.554 "zcopy": true, 00:14:22.554 "get_zone_info": false, 00:14:22.554 "zone_management": false, 00:14:22.554 "zone_append": false, 00:14:22.554 "compare": false, 00:14:22.554 "compare_and_write": false, 00:14:22.554 "abort": true, 00:14:22.554 "seek_hole": false, 00:14:22.554 "seek_data": false, 00:14:22.554 "copy": true, 00:14:22.554 "nvme_iov_md": false 00:14:22.554 }, 00:14:22.554 "memory_domains": [ 00:14:22.554 { 00:14:22.554 "dma_device_id": "system", 00:14:22.554 "dma_device_type": 1 00:14:22.554 }, 00:14:22.554 { 00:14:22.554 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:22.554 "dma_device_type": 2 00:14:22.554 } 00:14:22.554 ], 00:14:22.554 "driver_specific": {} 00:14:22.554 } 00:14:22.554 ] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:22.554 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.555 [2024-09-29 16:52:44.124471] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:22.555 [2024-09-29 16:52:44.124574] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:22.555 [2024-09-29 16:52:44.124613] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:22.555 [2024-09-29 16:52:44.126431] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:22.555 [2024-09-29 16:52:44.126532] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:22.555 "name": "Existed_Raid", 00:14:22.555 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:22.555 "strip_size_kb": 64, 00:14:22.555 "state": "configuring", 00:14:22.555 "raid_level": "raid5f", 00:14:22.555 "superblock": true, 00:14:22.555 "num_base_bdevs": 4, 00:14:22.555 "num_base_bdevs_discovered": 3, 00:14:22.555 "num_base_bdevs_operational": 4, 00:14:22.555 "base_bdevs_list": [ 00:14:22.555 { 00:14:22.555 "name": "BaseBdev1", 00:14:22.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:22.555 "is_configured": false, 00:14:22.555 "data_offset": 0, 00:14:22.555 "data_size": 0 00:14:22.555 }, 00:14:22.555 { 00:14:22.555 "name": "BaseBdev2", 00:14:22.555 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:22.555 "is_configured": true, 00:14:22.555 "data_offset": 2048, 00:14:22.555 "data_size": 63488 00:14:22.555 }, 00:14:22.555 { 00:14:22.555 "name": "BaseBdev3", 00:14:22.555 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:22.555 "is_configured": true, 00:14:22.555 "data_offset": 2048, 00:14:22.555 "data_size": 63488 00:14:22.555 }, 00:14:22.555 { 00:14:22.555 "name": "BaseBdev4", 00:14:22.555 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:22.555 "is_configured": true, 00:14:22.555 "data_offset": 2048, 00:14:22.555 "data_size": 63488 00:14:22.555 } 00:14:22.555 ] 00:14:22.555 }' 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:22.555 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.124 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:23.124 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.124 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.125 [2024-09-29 16:52:44.571704] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.125 "name": "Existed_Raid", 00:14:23.125 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:23.125 "strip_size_kb": 64, 00:14:23.125 "state": "configuring", 00:14:23.125 "raid_level": "raid5f", 00:14:23.125 "superblock": true, 00:14:23.125 "num_base_bdevs": 4, 00:14:23.125 "num_base_bdevs_discovered": 2, 00:14:23.125 "num_base_bdevs_operational": 4, 00:14:23.125 "base_bdevs_list": [ 00:14:23.125 { 00:14:23.125 "name": "BaseBdev1", 00:14:23.125 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:23.125 "is_configured": false, 00:14:23.125 "data_offset": 0, 00:14:23.125 "data_size": 0 00:14:23.125 }, 00:14:23.125 { 00:14:23.125 "name": null, 00:14:23.125 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:23.125 "is_configured": false, 00:14:23.125 "data_offset": 0, 00:14:23.125 "data_size": 63488 00:14:23.125 }, 00:14:23.125 { 00:14:23.125 "name": "BaseBdev3", 00:14:23.125 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:23.125 "is_configured": true, 00:14:23.125 "data_offset": 2048, 00:14:23.125 "data_size": 63488 00:14:23.125 }, 00:14:23.125 { 00:14:23.125 "name": "BaseBdev4", 00:14:23.125 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:23.125 "is_configured": true, 00:14:23.125 "data_offset": 2048, 00:14:23.125 "data_size": 63488 00:14:23.125 } 00:14:23.125 ] 00:14:23.125 }' 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.125 16:52:44 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.384 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.384 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.384 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.384 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:23.644 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.644 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:23.644 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:23.644 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.644 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.645 [2024-09-29 16:52:45.114027] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:23.645 BaseBdev1 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.645 [ 00:14:23.645 { 00:14:23.645 "name": "BaseBdev1", 00:14:23.645 "aliases": [ 00:14:23.645 "3017d3f9-f2f0-4785-8ad9-b2626b4312bd" 00:14:23.645 ], 00:14:23.645 "product_name": "Malloc disk", 00:14:23.645 "block_size": 512, 00:14:23.645 "num_blocks": 65536, 00:14:23.645 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:23.645 "assigned_rate_limits": { 00:14:23.645 "rw_ios_per_sec": 0, 00:14:23.645 "rw_mbytes_per_sec": 0, 00:14:23.645 "r_mbytes_per_sec": 0, 00:14:23.645 "w_mbytes_per_sec": 0 00:14:23.645 }, 00:14:23.645 "claimed": true, 00:14:23.645 "claim_type": "exclusive_write", 00:14:23.645 "zoned": false, 00:14:23.645 "supported_io_types": { 00:14:23.645 "read": true, 00:14:23.645 "write": true, 00:14:23.645 "unmap": true, 00:14:23.645 "flush": true, 00:14:23.645 "reset": true, 00:14:23.645 "nvme_admin": false, 00:14:23.645 "nvme_io": false, 00:14:23.645 "nvme_io_md": false, 00:14:23.645 "write_zeroes": true, 00:14:23.645 "zcopy": true, 00:14:23.645 "get_zone_info": false, 00:14:23.645 "zone_management": false, 00:14:23.645 "zone_append": false, 00:14:23.645 "compare": false, 00:14:23.645 "compare_and_write": false, 00:14:23.645 "abort": true, 00:14:23.645 "seek_hole": false, 00:14:23.645 "seek_data": false, 00:14:23.645 "copy": true, 00:14:23.645 "nvme_iov_md": false 00:14:23.645 }, 00:14:23.645 "memory_domains": [ 00:14:23.645 { 00:14:23.645 "dma_device_id": "system", 00:14:23.645 "dma_device_type": 1 00:14:23.645 }, 00:14:23.645 { 00:14:23.645 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:23.645 "dma_device_type": 2 00:14:23.645 } 00:14:23.645 ], 00:14:23.645 "driver_specific": {} 00:14:23.645 } 00:14:23.645 ] 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:23.645 "name": "Existed_Raid", 00:14:23.645 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:23.645 "strip_size_kb": 64, 00:14:23.645 "state": "configuring", 00:14:23.645 "raid_level": "raid5f", 00:14:23.645 "superblock": true, 00:14:23.645 "num_base_bdevs": 4, 00:14:23.645 "num_base_bdevs_discovered": 3, 00:14:23.645 "num_base_bdevs_operational": 4, 00:14:23.645 "base_bdevs_list": [ 00:14:23.645 { 00:14:23.645 "name": "BaseBdev1", 00:14:23.645 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:23.645 "is_configured": true, 00:14:23.645 "data_offset": 2048, 00:14:23.645 "data_size": 63488 00:14:23.645 }, 00:14:23.645 { 00:14:23.645 "name": null, 00:14:23.645 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:23.645 "is_configured": false, 00:14:23.645 "data_offset": 0, 00:14:23.645 "data_size": 63488 00:14:23.645 }, 00:14:23.645 { 00:14:23.645 "name": "BaseBdev3", 00:14:23.645 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:23.645 "is_configured": true, 00:14:23.645 "data_offset": 2048, 00:14:23.645 "data_size": 63488 00:14:23.645 }, 00:14:23.645 { 00:14:23.645 "name": "BaseBdev4", 00:14:23.645 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:23.645 "is_configured": true, 00:14:23.645 "data_offset": 2048, 00:14:23.645 "data_size": 63488 00:14:23.645 } 00:14:23.645 ] 00:14:23.645 }' 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:23.645 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.215 [2024-09-29 16:52:45.641139] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.215 "name": "Existed_Raid", 00:14:24.215 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:24.215 "strip_size_kb": 64, 00:14:24.215 "state": "configuring", 00:14:24.215 "raid_level": "raid5f", 00:14:24.215 "superblock": true, 00:14:24.215 "num_base_bdevs": 4, 00:14:24.215 "num_base_bdevs_discovered": 2, 00:14:24.215 "num_base_bdevs_operational": 4, 00:14:24.215 "base_bdevs_list": [ 00:14:24.215 { 00:14:24.215 "name": "BaseBdev1", 00:14:24.215 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:24.215 "is_configured": true, 00:14:24.215 "data_offset": 2048, 00:14:24.215 "data_size": 63488 00:14:24.215 }, 00:14:24.215 { 00:14:24.215 "name": null, 00:14:24.215 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:24.215 "is_configured": false, 00:14:24.215 "data_offset": 0, 00:14:24.215 "data_size": 63488 00:14:24.215 }, 00:14:24.215 { 00:14:24.215 "name": null, 00:14:24.215 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:24.215 "is_configured": false, 00:14:24.215 "data_offset": 0, 00:14:24.215 "data_size": 63488 00:14:24.215 }, 00:14:24.215 { 00:14:24.215 "name": "BaseBdev4", 00:14:24.215 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:24.215 "is_configured": true, 00:14:24.215 "data_offset": 2048, 00:14:24.215 "data_size": 63488 00:14:24.215 } 00:14:24.215 ] 00:14:24.215 }' 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.215 16:52:45 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.475 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.476 [2024-09-29 16:52:46.140327] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.476 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.735 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.735 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.735 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:24.735 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.736 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.736 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.736 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.736 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:24.736 "name": "Existed_Raid", 00:14:24.736 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:24.736 "strip_size_kb": 64, 00:14:24.736 "state": "configuring", 00:14:24.736 "raid_level": "raid5f", 00:14:24.736 "superblock": true, 00:14:24.736 "num_base_bdevs": 4, 00:14:24.736 "num_base_bdevs_discovered": 3, 00:14:24.736 "num_base_bdevs_operational": 4, 00:14:24.736 "base_bdevs_list": [ 00:14:24.736 { 00:14:24.736 "name": "BaseBdev1", 00:14:24.736 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:24.736 "is_configured": true, 00:14:24.736 "data_offset": 2048, 00:14:24.736 "data_size": 63488 00:14:24.736 }, 00:14:24.736 { 00:14:24.736 "name": null, 00:14:24.736 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:24.736 "is_configured": false, 00:14:24.736 "data_offset": 0, 00:14:24.736 "data_size": 63488 00:14:24.736 }, 00:14:24.736 { 00:14:24.736 "name": "BaseBdev3", 00:14:24.736 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:24.736 "is_configured": true, 00:14:24.736 "data_offset": 2048, 00:14:24.736 "data_size": 63488 00:14:24.736 }, 00:14:24.736 { 00:14:24.736 "name": "BaseBdev4", 00:14:24.736 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:24.736 "is_configured": true, 00:14:24.736 "data_offset": 2048, 00:14:24.736 "data_size": 63488 00:14:24.736 } 00:14:24.736 ] 00:14:24.736 }' 00:14:24.736 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:24.736 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.996 [2024-09-29 16:52:46.623487] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:24.996 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.256 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:25.256 "name": "Existed_Raid", 00:14:25.256 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:25.256 "strip_size_kb": 64, 00:14:25.256 "state": "configuring", 00:14:25.256 "raid_level": "raid5f", 00:14:25.256 "superblock": true, 00:14:25.256 "num_base_bdevs": 4, 00:14:25.256 "num_base_bdevs_discovered": 2, 00:14:25.256 "num_base_bdevs_operational": 4, 00:14:25.256 "base_bdevs_list": [ 00:14:25.256 { 00:14:25.256 "name": null, 00:14:25.256 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:25.256 "is_configured": false, 00:14:25.256 "data_offset": 0, 00:14:25.256 "data_size": 63488 00:14:25.256 }, 00:14:25.256 { 00:14:25.256 "name": null, 00:14:25.256 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:25.256 "is_configured": false, 00:14:25.256 "data_offset": 0, 00:14:25.256 "data_size": 63488 00:14:25.256 }, 00:14:25.256 { 00:14:25.256 "name": "BaseBdev3", 00:14:25.256 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:25.256 "is_configured": true, 00:14:25.256 "data_offset": 2048, 00:14:25.256 "data_size": 63488 00:14:25.256 }, 00:14:25.256 { 00:14:25.256 "name": "BaseBdev4", 00:14:25.256 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:25.256 "is_configured": true, 00:14:25.256 "data_offset": 2048, 00:14:25.256 "data_size": 63488 00:14:25.256 } 00:14:25.256 ] 00:14:25.256 }' 00:14:25.256 16:52:46 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:25.256 16:52:46 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.516 [2024-09-29 16:52:47.121208] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:25.516 "name": "Existed_Raid", 00:14:25.516 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:25.516 "strip_size_kb": 64, 00:14:25.516 "state": "configuring", 00:14:25.516 "raid_level": "raid5f", 00:14:25.516 "superblock": true, 00:14:25.516 "num_base_bdevs": 4, 00:14:25.516 "num_base_bdevs_discovered": 3, 00:14:25.516 "num_base_bdevs_operational": 4, 00:14:25.516 "base_bdevs_list": [ 00:14:25.516 { 00:14:25.516 "name": null, 00:14:25.516 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:25.516 "is_configured": false, 00:14:25.516 "data_offset": 0, 00:14:25.516 "data_size": 63488 00:14:25.516 }, 00:14:25.516 { 00:14:25.516 "name": "BaseBdev2", 00:14:25.516 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:25.516 "is_configured": true, 00:14:25.516 "data_offset": 2048, 00:14:25.516 "data_size": 63488 00:14:25.516 }, 00:14:25.516 { 00:14:25.516 "name": "BaseBdev3", 00:14:25.516 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:25.516 "is_configured": true, 00:14:25.516 "data_offset": 2048, 00:14:25.516 "data_size": 63488 00:14:25.516 }, 00:14:25.516 { 00:14:25.516 "name": "BaseBdev4", 00:14:25.516 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:25.516 "is_configured": true, 00:14:25.516 "data_offset": 2048, 00:14:25.516 "data_size": 63488 00:14:25.516 } 00:14:25.516 ] 00:14:25.516 }' 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:25.516 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 3017d3f9-f2f0-4785-8ad9-b2626b4312bd 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.086 NewBaseBdev 00:14:26.086 [2024-09-29 16:52:47.663037] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:26.086 [2024-09-29 16:52:47.663227] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:26.086 [2024-09-29 16:52:47.663240] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:26.086 [2024-09-29 16:52:47.663501] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:26.086 [2024-09-29 16:52:47.664007] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:26.086 [2024-09-29 16:52:47.664028] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:26.086 [2024-09-29 16:52:47.664126] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.086 [ 00:14:26.086 { 00:14:26.086 "name": "NewBaseBdev", 00:14:26.086 "aliases": [ 00:14:26.086 "3017d3f9-f2f0-4785-8ad9-b2626b4312bd" 00:14:26.086 ], 00:14:26.086 "product_name": "Malloc disk", 00:14:26.086 "block_size": 512, 00:14:26.086 "num_blocks": 65536, 00:14:26.086 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:26.086 "assigned_rate_limits": { 00:14:26.086 "rw_ios_per_sec": 0, 00:14:26.086 "rw_mbytes_per_sec": 0, 00:14:26.086 "r_mbytes_per_sec": 0, 00:14:26.086 "w_mbytes_per_sec": 0 00:14:26.086 }, 00:14:26.086 "claimed": true, 00:14:26.086 "claim_type": "exclusive_write", 00:14:26.086 "zoned": false, 00:14:26.086 "supported_io_types": { 00:14:26.086 "read": true, 00:14:26.086 "write": true, 00:14:26.086 "unmap": true, 00:14:26.086 "flush": true, 00:14:26.086 "reset": true, 00:14:26.086 "nvme_admin": false, 00:14:26.086 "nvme_io": false, 00:14:26.086 "nvme_io_md": false, 00:14:26.086 "write_zeroes": true, 00:14:26.086 "zcopy": true, 00:14:26.086 "get_zone_info": false, 00:14:26.086 "zone_management": false, 00:14:26.086 "zone_append": false, 00:14:26.086 "compare": false, 00:14:26.086 "compare_and_write": false, 00:14:26.086 "abort": true, 00:14:26.086 "seek_hole": false, 00:14:26.086 "seek_data": false, 00:14:26.086 "copy": true, 00:14:26.086 "nvme_iov_md": false 00:14:26.086 }, 00:14:26.086 "memory_domains": [ 00:14:26.086 { 00:14:26.086 "dma_device_id": "system", 00:14:26.086 "dma_device_type": 1 00:14:26.086 }, 00:14:26.086 { 00:14:26.086 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:26.086 "dma_device_type": 2 00:14:26.086 } 00:14:26.086 ], 00:14:26.086 "driver_specific": {} 00:14:26.086 } 00:14:26.086 ] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:26.086 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.087 "name": "Existed_Raid", 00:14:26.087 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:26.087 "strip_size_kb": 64, 00:14:26.087 "state": "online", 00:14:26.087 "raid_level": "raid5f", 00:14:26.087 "superblock": true, 00:14:26.087 "num_base_bdevs": 4, 00:14:26.087 "num_base_bdevs_discovered": 4, 00:14:26.087 "num_base_bdevs_operational": 4, 00:14:26.087 "base_bdevs_list": [ 00:14:26.087 { 00:14:26.087 "name": "NewBaseBdev", 00:14:26.087 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:26.087 "is_configured": true, 00:14:26.087 "data_offset": 2048, 00:14:26.087 "data_size": 63488 00:14:26.087 }, 00:14:26.087 { 00:14:26.087 "name": "BaseBdev2", 00:14:26.087 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:26.087 "is_configured": true, 00:14:26.087 "data_offset": 2048, 00:14:26.087 "data_size": 63488 00:14:26.087 }, 00:14:26.087 { 00:14:26.087 "name": "BaseBdev3", 00:14:26.087 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:26.087 "is_configured": true, 00:14:26.087 "data_offset": 2048, 00:14:26.087 "data_size": 63488 00:14:26.087 }, 00:14:26.087 { 00:14:26.087 "name": "BaseBdev4", 00:14:26.087 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:26.087 "is_configured": true, 00:14:26.087 "data_offset": 2048, 00:14:26.087 "data_size": 63488 00:14:26.087 } 00:14:26.087 ] 00:14:26.087 }' 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.087 16:52:47 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:26.656 [2024-09-29 16:52:48.126435] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:26.656 "name": "Existed_Raid", 00:14:26.656 "aliases": [ 00:14:26.656 "728ed291-c092-4a13-af51-dca7945a4f0b" 00:14:26.656 ], 00:14:26.656 "product_name": "Raid Volume", 00:14:26.656 "block_size": 512, 00:14:26.656 "num_blocks": 190464, 00:14:26.656 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:26.656 "assigned_rate_limits": { 00:14:26.656 "rw_ios_per_sec": 0, 00:14:26.656 "rw_mbytes_per_sec": 0, 00:14:26.656 "r_mbytes_per_sec": 0, 00:14:26.656 "w_mbytes_per_sec": 0 00:14:26.656 }, 00:14:26.656 "claimed": false, 00:14:26.656 "zoned": false, 00:14:26.656 "supported_io_types": { 00:14:26.656 "read": true, 00:14:26.656 "write": true, 00:14:26.656 "unmap": false, 00:14:26.656 "flush": false, 00:14:26.656 "reset": true, 00:14:26.656 "nvme_admin": false, 00:14:26.656 "nvme_io": false, 00:14:26.656 "nvme_io_md": false, 00:14:26.656 "write_zeroes": true, 00:14:26.656 "zcopy": false, 00:14:26.656 "get_zone_info": false, 00:14:26.656 "zone_management": false, 00:14:26.656 "zone_append": false, 00:14:26.656 "compare": false, 00:14:26.656 "compare_and_write": false, 00:14:26.656 "abort": false, 00:14:26.656 "seek_hole": false, 00:14:26.656 "seek_data": false, 00:14:26.656 "copy": false, 00:14:26.656 "nvme_iov_md": false 00:14:26.656 }, 00:14:26.656 "driver_specific": { 00:14:26.656 "raid": { 00:14:26.656 "uuid": "728ed291-c092-4a13-af51-dca7945a4f0b", 00:14:26.656 "strip_size_kb": 64, 00:14:26.656 "state": "online", 00:14:26.656 "raid_level": "raid5f", 00:14:26.656 "superblock": true, 00:14:26.656 "num_base_bdevs": 4, 00:14:26.656 "num_base_bdevs_discovered": 4, 00:14:26.656 "num_base_bdevs_operational": 4, 00:14:26.656 "base_bdevs_list": [ 00:14:26.656 { 00:14:26.656 "name": "NewBaseBdev", 00:14:26.656 "uuid": "3017d3f9-f2f0-4785-8ad9-b2626b4312bd", 00:14:26.656 "is_configured": true, 00:14:26.656 "data_offset": 2048, 00:14:26.656 "data_size": 63488 00:14:26.656 }, 00:14:26.656 { 00:14:26.656 "name": "BaseBdev2", 00:14:26.656 "uuid": "ebca5f27-a2ae-4cb8-a831-dc724ffc2947", 00:14:26.656 "is_configured": true, 00:14:26.656 "data_offset": 2048, 00:14:26.656 "data_size": 63488 00:14:26.656 }, 00:14:26.656 { 00:14:26.656 "name": "BaseBdev3", 00:14:26.656 "uuid": "e4ceebd9-e2ec-4cff-8499-14f1bae6deb5", 00:14:26.656 "is_configured": true, 00:14:26.656 "data_offset": 2048, 00:14:26.656 "data_size": 63488 00:14:26.656 }, 00:14:26.656 { 00:14:26.656 "name": "BaseBdev4", 00:14:26.656 "uuid": "9ea5dd48-9810-425a-ac11-f4cc0cc80999", 00:14:26.656 "is_configured": true, 00:14:26.656 "data_offset": 2048, 00:14:26.656 "data_size": 63488 00:14:26.656 } 00:14:26.656 ] 00:14:26.656 } 00:14:26.656 } 00:14:26.656 }' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:26.656 BaseBdev2 00:14:26.656 BaseBdev3 00:14:26.656 BaseBdev4' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.656 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:26.916 [2024-09-29 16:52:48.433811] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:26.916 [2024-09-29 16:52:48.433837] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:26.916 [2024-09-29 16:52:48.433899] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:26.916 [2024-09-29 16:52:48.434141] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:26.916 [2024-09-29 16:52:48.434152] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 93543 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 93543 ']' 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 93543 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 93543 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:26.916 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:26.917 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 93543' 00:14:26.917 killing process with pid 93543 00:14:26.917 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 93543 00:14:26.917 [2024-09-29 16:52:48.473094] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:26.917 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 93543 00:14:26.917 [2024-09-29 16:52:48.514193] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:27.177 16:52:48 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:27.177 00:14:27.177 real 0m9.786s 00:14:27.177 user 0m16.770s 00:14:27.177 sys 0m2.091s 00:14:27.177 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:27.177 16:52:48 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.177 ************************************ 00:14:27.177 END TEST raid5f_state_function_test_sb 00:14:27.177 ************************************ 00:14:27.177 16:52:48 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:14:27.177 16:52:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:14:27.177 16:52:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:27.177 16:52:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:27.177 ************************************ 00:14:27.177 START TEST raid5f_superblock_test 00:14:27.177 ************************************ 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 4 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=94197 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 94197 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 94197 ']' 00:14:27.177 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:27.177 16:52:48 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.444 [2024-09-29 16:52:48.936658] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:14:27.444 [2024-09-29 16:52:48.936939] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94197 ] 00:14:27.444 [2024-09-29 16:52:49.082580] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:27.749 [2024-09-29 16:52:49.128928] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:27.749 [2024-09-29 16:52:49.172081] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:27.749 [2024-09-29 16:52:49.172119] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.325 malloc1 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.325 [2024-09-29 16:52:49.766772] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:28.325 [2024-09-29 16:52:49.766875] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.325 [2024-09-29 16:52:49.766907] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:28.325 [2024-09-29 16:52:49.766939] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.325 [2024-09-29 16:52:49.769107] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.325 [2024-09-29 16:52:49.769196] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:28.325 pt1 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.325 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 malloc2 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 [2024-09-29 16:52:49.816712] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:28.326 [2024-09-29 16:52:49.816833] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.326 [2024-09-29 16:52:49.816868] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:28.326 [2024-09-29 16:52:49.816892] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.326 pt2 00:14:28.326 [2024-09-29 16:52:49.821103] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.326 [2024-09-29 16:52:49.821156] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 malloc3 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 [2024-09-29 16:52:49.847264] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:28.326 [2024-09-29 16:52:49.847375] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.326 [2024-09-29 16:52:49.847409] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:28.326 [2024-09-29 16:52:49.847438] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.326 [2024-09-29 16:52:49.849519] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.326 [2024-09-29 16:52:49.849610] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:28.326 pt3 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 malloc4 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 [2024-09-29 16:52:49.880040] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:28.326 [2024-09-29 16:52:49.880128] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.326 [2024-09-29 16:52:49.880161] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:28.326 [2024-09-29 16:52:49.880199] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.326 [2024-09-29 16:52:49.882295] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.326 [2024-09-29 16:52:49.882363] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:28.326 pt4 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 [2024-09-29 16:52:49.892066] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:28.326 [2024-09-29 16:52:49.893900] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:28.326 [2024-09-29 16:52:49.894000] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:28.326 [2024-09-29 16:52:49.894077] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:28.326 [2024-09-29 16:52:49.894290] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:28.326 [2024-09-29 16:52:49.894338] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:28.326 [2024-09-29 16:52:49.894587] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:28.326 [2024-09-29 16:52:49.895073] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:28.326 [2024-09-29 16:52:49.895086] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:28.326 [2024-09-29 16:52:49.895217] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.326 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:28.326 "name": "raid_bdev1", 00:14:28.326 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:28.326 "strip_size_kb": 64, 00:14:28.326 "state": "online", 00:14:28.326 "raid_level": "raid5f", 00:14:28.326 "superblock": true, 00:14:28.326 "num_base_bdevs": 4, 00:14:28.326 "num_base_bdevs_discovered": 4, 00:14:28.326 "num_base_bdevs_operational": 4, 00:14:28.326 "base_bdevs_list": [ 00:14:28.326 { 00:14:28.326 "name": "pt1", 00:14:28.326 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:28.326 "is_configured": true, 00:14:28.326 "data_offset": 2048, 00:14:28.326 "data_size": 63488 00:14:28.326 }, 00:14:28.326 { 00:14:28.326 "name": "pt2", 00:14:28.326 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:28.326 "is_configured": true, 00:14:28.326 "data_offset": 2048, 00:14:28.326 "data_size": 63488 00:14:28.326 }, 00:14:28.326 { 00:14:28.326 "name": "pt3", 00:14:28.326 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:28.326 "is_configured": true, 00:14:28.326 "data_offset": 2048, 00:14:28.326 "data_size": 63488 00:14:28.326 }, 00:14:28.326 { 00:14:28.326 "name": "pt4", 00:14:28.326 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:28.327 "is_configured": true, 00:14:28.327 "data_offset": 2048, 00:14:28.327 "data_size": 63488 00:14:28.327 } 00:14:28.327 ] 00:14:28.327 }' 00:14:28.327 16:52:49 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:28.327 16:52:49 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.895 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.896 [2024-09-29 16:52:50.336338] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:28.896 "name": "raid_bdev1", 00:14:28.896 "aliases": [ 00:14:28.896 "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd" 00:14:28.896 ], 00:14:28.896 "product_name": "Raid Volume", 00:14:28.896 "block_size": 512, 00:14:28.896 "num_blocks": 190464, 00:14:28.896 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:28.896 "assigned_rate_limits": { 00:14:28.896 "rw_ios_per_sec": 0, 00:14:28.896 "rw_mbytes_per_sec": 0, 00:14:28.896 "r_mbytes_per_sec": 0, 00:14:28.896 "w_mbytes_per_sec": 0 00:14:28.896 }, 00:14:28.896 "claimed": false, 00:14:28.896 "zoned": false, 00:14:28.896 "supported_io_types": { 00:14:28.896 "read": true, 00:14:28.896 "write": true, 00:14:28.896 "unmap": false, 00:14:28.896 "flush": false, 00:14:28.896 "reset": true, 00:14:28.896 "nvme_admin": false, 00:14:28.896 "nvme_io": false, 00:14:28.896 "nvme_io_md": false, 00:14:28.896 "write_zeroes": true, 00:14:28.896 "zcopy": false, 00:14:28.896 "get_zone_info": false, 00:14:28.896 "zone_management": false, 00:14:28.896 "zone_append": false, 00:14:28.896 "compare": false, 00:14:28.896 "compare_and_write": false, 00:14:28.896 "abort": false, 00:14:28.896 "seek_hole": false, 00:14:28.896 "seek_data": false, 00:14:28.896 "copy": false, 00:14:28.896 "nvme_iov_md": false 00:14:28.896 }, 00:14:28.896 "driver_specific": { 00:14:28.896 "raid": { 00:14:28.896 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:28.896 "strip_size_kb": 64, 00:14:28.896 "state": "online", 00:14:28.896 "raid_level": "raid5f", 00:14:28.896 "superblock": true, 00:14:28.896 "num_base_bdevs": 4, 00:14:28.896 "num_base_bdevs_discovered": 4, 00:14:28.896 "num_base_bdevs_operational": 4, 00:14:28.896 "base_bdevs_list": [ 00:14:28.896 { 00:14:28.896 "name": "pt1", 00:14:28.896 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:28.896 "is_configured": true, 00:14:28.896 "data_offset": 2048, 00:14:28.896 "data_size": 63488 00:14:28.896 }, 00:14:28.896 { 00:14:28.896 "name": "pt2", 00:14:28.896 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:28.896 "is_configured": true, 00:14:28.896 "data_offset": 2048, 00:14:28.896 "data_size": 63488 00:14:28.896 }, 00:14:28.896 { 00:14:28.896 "name": "pt3", 00:14:28.896 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:28.896 "is_configured": true, 00:14:28.896 "data_offset": 2048, 00:14:28.896 "data_size": 63488 00:14:28.896 }, 00:14:28.896 { 00:14:28.896 "name": "pt4", 00:14:28.896 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:28.896 "is_configured": true, 00:14:28.896 "data_offset": 2048, 00:14:28.896 "data_size": 63488 00:14:28.896 } 00:14:28.896 ] 00:14:28.896 } 00:14:28.896 } 00:14:28.896 }' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:28.896 pt2 00:14:28.896 pt3 00:14:28.896 pt4' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:28.896 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.156 [2024-09-29 16:52:50.679809] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=8255bd53-cbbc-4092-9fdd-c69fd5ade7cd 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 8255bd53-cbbc-4092-9fdd-c69fd5ade7cd ']' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.156 [2024-09-29 16:52:50.727577] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:29.156 [2024-09-29 16:52:50.727647] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:29.156 [2024-09-29 16:52:50.727754] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:29.156 [2024-09-29 16:52:50.727869] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:29.156 [2024-09-29 16:52:50.727924] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:29.156 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.157 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.416 [2024-09-29 16:52:50.895326] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:29.416 [2024-09-29 16:52:50.897203] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:29.416 [2024-09-29 16:52:50.897246] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:29.416 [2024-09-29 16:52:50.897280] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:14:29.416 [2024-09-29 16:52:50.897324] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:29.416 [2024-09-29 16:52:50.897363] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:29.416 [2024-09-29 16:52:50.897381] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:29.416 [2024-09-29 16:52:50.897396] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:14:29.416 [2024-09-29 16:52:50.897409] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:29.416 [2024-09-29 16:52:50.897419] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:14:29.416 request: 00:14:29.416 { 00:14:29.416 "name": "raid_bdev1", 00:14:29.416 "raid_level": "raid5f", 00:14:29.416 "base_bdevs": [ 00:14:29.416 "malloc1", 00:14:29.416 "malloc2", 00:14:29.416 "malloc3", 00:14:29.416 "malloc4" 00:14:29.416 ], 00:14:29.416 "strip_size_kb": 64, 00:14:29.416 "superblock": false, 00:14:29.416 "method": "bdev_raid_create", 00:14:29.416 "req_id": 1 00:14:29.416 } 00:14:29.416 Got JSON-RPC error response 00:14:29.416 response: 00:14:29.416 { 00:14:29.416 "code": -17, 00:14:29.416 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:29.416 } 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.416 [2024-09-29 16:52:50.959229] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:29.416 [2024-09-29 16:52:50.959312] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:29.416 [2024-09-29 16:52:50.959351] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:29.416 [2024-09-29 16:52:50.959401] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:29.416 [2024-09-29 16:52:50.961503] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:29.416 [2024-09-29 16:52:50.961572] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:29.416 [2024-09-29 16:52:50.961656] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:29.416 [2024-09-29 16:52:50.961733] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:29.416 pt1 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.416 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.417 16:52:50 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.417 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.417 "name": "raid_bdev1", 00:14:29.417 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:29.417 "strip_size_kb": 64, 00:14:29.417 "state": "configuring", 00:14:29.417 "raid_level": "raid5f", 00:14:29.417 "superblock": true, 00:14:29.417 "num_base_bdevs": 4, 00:14:29.417 "num_base_bdevs_discovered": 1, 00:14:29.417 "num_base_bdevs_operational": 4, 00:14:29.417 "base_bdevs_list": [ 00:14:29.417 { 00:14:29.417 "name": "pt1", 00:14:29.417 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:29.417 "is_configured": true, 00:14:29.417 "data_offset": 2048, 00:14:29.417 "data_size": 63488 00:14:29.417 }, 00:14:29.417 { 00:14:29.417 "name": null, 00:14:29.417 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:29.417 "is_configured": false, 00:14:29.417 "data_offset": 2048, 00:14:29.417 "data_size": 63488 00:14:29.417 }, 00:14:29.417 { 00:14:29.417 "name": null, 00:14:29.417 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:29.417 "is_configured": false, 00:14:29.417 "data_offset": 2048, 00:14:29.417 "data_size": 63488 00:14:29.417 }, 00:14:29.417 { 00:14:29.417 "name": null, 00:14:29.417 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:29.417 "is_configured": false, 00:14:29.417 "data_offset": 2048, 00:14:29.417 "data_size": 63488 00:14:29.417 } 00:14:29.417 ] 00:14:29.417 }' 00:14:29.417 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.417 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.985 [2024-09-29 16:52:51.382476] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:29.985 [2024-09-29 16:52:51.382558] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:29.985 [2024-09-29 16:52:51.382578] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:29.985 [2024-09-29 16:52:51.382587] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:29.985 [2024-09-29 16:52:51.382926] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:29.985 [2024-09-29 16:52:51.382944] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:29.985 [2024-09-29 16:52:51.382995] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:29.985 [2024-09-29 16:52:51.383019] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:29.985 pt2 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.985 [2024-09-29 16:52:51.390501] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:29.985 "name": "raid_bdev1", 00:14:29.985 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:29.985 "strip_size_kb": 64, 00:14:29.985 "state": "configuring", 00:14:29.985 "raid_level": "raid5f", 00:14:29.985 "superblock": true, 00:14:29.985 "num_base_bdevs": 4, 00:14:29.985 "num_base_bdevs_discovered": 1, 00:14:29.985 "num_base_bdevs_operational": 4, 00:14:29.985 "base_bdevs_list": [ 00:14:29.985 { 00:14:29.985 "name": "pt1", 00:14:29.985 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:29.985 "is_configured": true, 00:14:29.985 "data_offset": 2048, 00:14:29.985 "data_size": 63488 00:14:29.985 }, 00:14:29.985 { 00:14:29.985 "name": null, 00:14:29.985 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:29.985 "is_configured": false, 00:14:29.985 "data_offset": 0, 00:14:29.985 "data_size": 63488 00:14:29.985 }, 00:14:29.985 { 00:14:29.985 "name": null, 00:14:29.985 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:29.985 "is_configured": false, 00:14:29.985 "data_offset": 2048, 00:14:29.985 "data_size": 63488 00:14:29.985 }, 00:14:29.985 { 00:14:29.985 "name": null, 00:14:29.985 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:29.985 "is_configured": false, 00:14:29.985 "data_offset": 2048, 00:14:29.985 "data_size": 63488 00:14:29.985 } 00:14:29.985 ] 00:14:29.985 }' 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:29.985 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.245 [2024-09-29 16:52:51.869709] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:30.245 [2024-09-29 16:52:51.869823] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.245 [2024-09-29 16:52:51.869857] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:14:30.245 [2024-09-29 16:52:51.869894] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.245 [2024-09-29 16:52:51.870252] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.245 [2024-09-29 16:52:51.870311] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:30.245 [2024-09-29 16:52:51.870400] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:30.245 [2024-09-29 16:52:51.870448] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:30.245 pt2 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.245 [2024-09-29 16:52:51.881668] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:30.245 [2024-09-29 16:52:51.881767] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.245 [2024-09-29 16:52:51.881801] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:30.245 [2024-09-29 16:52:51.881852] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.245 [2024-09-29 16:52:51.882173] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.245 [2024-09-29 16:52:51.882230] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:30.245 [2024-09-29 16:52:51.882325] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:30.245 [2024-09-29 16:52:51.882372] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:30.245 pt3 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.245 [2024-09-29 16:52:51.893688] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:30.245 [2024-09-29 16:52:51.893742] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:30.245 [2024-09-29 16:52:51.893771] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:14:30.245 [2024-09-29 16:52:51.893781] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:30.245 [2024-09-29 16:52:51.894046] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:30.245 [2024-09-29 16:52:51.894066] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:30.245 [2024-09-29 16:52:51.894112] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:30.245 [2024-09-29 16:52:51.894129] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:30.245 [2024-09-29 16:52:51.894230] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:30.245 [2024-09-29 16:52:51.894267] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:30.245 [2024-09-29 16:52:51.894491] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:30.245 [2024-09-29 16:52:51.894945] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:30.245 [2024-09-29 16:52:51.894963] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:14:30.245 [2024-09-29 16:52:51.895056] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:30.245 pt4 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.245 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.505 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.505 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:30.505 "name": "raid_bdev1", 00:14:30.505 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:30.505 "strip_size_kb": 64, 00:14:30.505 "state": "online", 00:14:30.505 "raid_level": "raid5f", 00:14:30.505 "superblock": true, 00:14:30.505 "num_base_bdevs": 4, 00:14:30.505 "num_base_bdevs_discovered": 4, 00:14:30.505 "num_base_bdevs_operational": 4, 00:14:30.505 "base_bdevs_list": [ 00:14:30.505 { 00:14:30.505 "name": "pt1", 00:14:30.505 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:30.505 "is_configured": true, 00:14:30.505 "data_offset": 2048, 00:14:30.505 "data_size": 63488 00:14:30.505 }, 00:14:30.505 { 00:14:30.505 "name": "pt2", 00:14:30.505 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:30.505 "is_configured": true, 00:14:30.505 "data_offset": 2048, 00:14:30.505 "data_size": 63488 00:14:30.505 }, 00:14:30.505 { 00:14:30.505 "name": "pt3", 00:14:30.505 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:30.505 "is_configured": true, 00:14:30.505 "data_offset": 2048, 00:14:30.505 "data_size": 63488 00:14:30.505 }, 00:14:30.505 { 00:14:30.505 "name": "pt4", 00:14:30.505 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:30.505 "is_configured": true, 00:14:30.505 "data_offset": 2048, 00:14:30.505 "data_size": 63488 00:14:30.505 } 00:14:30.505 ] 00:14:30.505 }' 00:14:30.505 16:52:51 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:30.505 16:52:51 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:30.764 [2024-09-29 16:52:52.369013] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.764 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:30.765 "name": "raid_bdev1", 00:14:30.765 "aliases": [ 00:14:30.765 "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd" 00:14:30.765 ], 00:14:30.765 "product_name": "Raid Volume", 00:14:30.765 "block_size": 512, 00:14:30.765 "num_blocks": 190464, 00:14:30.765 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:30.765 "assigned_rate_limits": { 00:14:30.765 "rw_ios_per_sec": 0, 00:14:30.765 "rw_mbytes_per_sec": 0, 00:14:30.765 "r_mbytes_per_sec": 0, 00:14:30.765 "w_mbytes_per_sec": 0 00:14:30.765 }, 00:14:30.765 "claimed": false, 00:14:30.765 "zoned": false, 00:14:30.765 "supported_io_types": { 00:14:30.765 "read": true, 00:14:30.765 "write": true, 00:14:30.765 "unmap": false, 00:14:30.765 "flush": false, 00:14:30.765 "reset": true, 00:14:30.765 "nvme_admin": false, 00:14:30.765 "nvme_io": false, 00:14:30.765 "nvme_io_md": false, 00:14:30.765 "write_zeroes": true, 00:14:30.765 "zcopy": false, 00:14:30.765 "get_zone_info": false, 00:14:30.765 "zone_management": false, 00:14:30.765 "zone_append": false, 00:14:30.765 "compare": false, 00:14:30.765 "compare_and_write": false, 00:14:30.765 "abort": false, 00:14:30.765 "seek_hole": false, 00:14:30.765 "seek_data": false, 00:14:30.765 "copy": false, 00:14:30.765 "nvme_iov_md": false 00:14:30.765 }, 00:14:30.765 "driver_specific": { 00:14:30.765 "raid": { 00:14:30.765 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:30.765 "strip_size_kb": 64, 00:14:30.765 "state": "online", 00:14:30.765 "raid_level": "raid5f", 00:14:30.765 "superblock": true, 00:14:30.765 "num_base_bdevs": 4, 00:14:30.765 "num_base_bdevs_discovered": 4, 00:14:30.765 "num_base_bdevs_operational": 4, 00:14:30.765 "base_bdevs_list": [ 00:14:30.765 { 00:14:30.765 "name": "pt1", 00:14:30.765 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:30.765 "is_configured": true, 00:14:30.765 "data_offset": 2048, 00:14:30.765 "data_size": 63488 00:14:30.765 }, 00:14:30.765 { 00:14:30.765 "name": "pt2", 00:14:30.765 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:30.765 "is_configured": true, 00:14:30.765 "data_offset": 2048, 00:14:30.765 "data_size": 63488 00:14:30.765 }, 00:14:30.765 { 00:14:30.765 "name": "pt3", 00:14:30.765 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:30.765 "is_configured": true, 00:14:30.765 "data_offset": 2048, 00:14:30.765 "data_size": 63488 00:14:30.765 }, 00:14:30.765 { 00:14:30.765 "name": "pt4", 00:14:30.765 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:30.765 "is_configured": true, 00:14:30.765 "data_offset": 2048, 00:14:30.765 "data_size": 63488 00:14:30.765 } 00:14:30.765 ] 00:14:30.765 } 00:14:30.765 } 00:14:30.765 }' 00:14:30.765 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:31.026 pt2 00:14:31.026 pt3 00:14:31.026 pt4' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:31.026 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:31.027 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.027 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.027 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:31.027 [2024-09-29 16:52:52.684450] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 8255bd53-cbbc-4092-9fdd-c69fd5ade7cd '!=' 8255bd53-cbbc-4092-9fdd-c69fd5ade7cd ']' 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.287 [2024-09-29 16:52:52.732228] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:31.287 "name": "raid_bdev1", 00:14:31.287 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:31.287 "strip_size_kb": 64, 00:14:31.287 "state": "online", 00:14:31.287 "raid_level": "raid5f", 00:14:31.287 "superblock": true, 00:14:31.287 "num_base_bdevs": 4, 00:14:31.287 "num_base_bdevs_discovered": 3, 00:14:31.287 "num_base_bdevs_operational": 3, 00:14:31.287 "base_bdevs_list": [ 00:14:31.287 { 00:14:31.287 "name": null, 00:14:31.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.287 "is_configured": false, 00:14:31.287 "data_offset": 0, 00:14:31.287 "data_size": 63488 00:14:31.287 }, 00:14:31.287 { 00:14:31.287 "name": "pt2", 00:14:31.287 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:31.287 "is_configured": true, 00:14:31.287 "data_offset": 2048, 00:14:31.287 "data_size": 63488 00:14:31.287 }, 00:14:31.287 { 00:14:31.287 "name": "pt3", 00:14:31.287 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:31.287 "is_configured": true, 00:14:31.287 "data_offset": 2048, 00:14:31.287 "data_size": 63488 00:14:31.287 }, 00:14:31.287 { 00:14:31.287 "name": "pt4", 00:14:31.287 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:31.287 "is_configured": true, 00:14:31.287 "data_offset": 2048, 00:14:31.287 "data_size": 63488 00:14:31.287 } 00:14:31.287 ] 00:14:31.287 }' 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:31.287 16:52:52 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.548 [2024-09-29 16:52:53.171452] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:31.548 [2024-09-29 16:52:53.171514] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:31.548 [2024-09-29 16:52:53.171587] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:31.548 [2024-09-29 16:52:53.171712] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:31.548 [2024-09-29 16:52:53.171771] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.548 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.808 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.808 [2024-09-29 16:52:53.271277] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:31.808 [2024-09-29 16:52:53.271324] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:31.808 [2024-09-29 16:52:53.271338] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:31.809 [2024-09-29 16:52:53.271348] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:31.809 [2024-09-29 16:52:53.273459] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:31.809 [2024-09-29 16:52:53.273497] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:31.809 [2024-09-29 16:52:53.273555] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:31.809 [2024-09-29 16:52:53.273594] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:31.809 pt2 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:31.809 "name": "raid_bdev1", 00:14:31.809 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:31.809 "strip_size_kb": 64, 00:14:31.809 "state": "configuring", 00:14:31.809 "raid_level": "raid5f", 00:14:31.809 "superblock": true, 00:14:31.809 "num_base_bdevs": 4, 00:14:31.809 "num_base_bdevs_discovered": 1, 00:14:31.809 "num_base_bdevs_operational": 3, 00:14:31.809 "base_bdevs_list": [ 00:14:31.809 { 00:14:31.809 "name": null, 00:14:31.809 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.809 "is_configured": false, 00:14:31.809 "data_offset": 2048, 00:14:31.809 "data_size": 63488 00:14:31.809 }, 00:14:31.809 { 00:14:31.809 "name": "pt2", 00:14:31.809 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:31.809 "is_configured": true, 00:14:31.809 "data_offset": 2048, 00:14:31.809 "data_size": 63488 00:14:31.809 }, 00:14:31.809 { 00:14:31.809 "name": null, 00:14:31.809 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:31.809 "is_configured": false, 00:14:31.809 "data_offset": 2048, 00:14:31.809 "data_size": 63488 00:14:31.809 }, 00:14:31.809 { 00:14:31.809 "name": null, 00:14:31.809 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:31.809 "is_configured": false, 00:14:31.809 "data_offset": 2048, 00:14:31.809 "data_size": 63488 00:14:31.809 } 00:14:31.809 ] 00:14:31.809 }' 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:31.809 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.068 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:32.068 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:32.068 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:32.068 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.068 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.068 [2024-09-29 16:52:53.738671] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:32.068 [2024-09-29 16:52:53.738774] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:32.068 [2024-09-29 16:52:53.738806] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:14:32.068 [2024-09-29 16:52:53.738837] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:32.068 [2024-09-29 16:52:53.739246] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:32.068 [2024-09-29 16:52:53.739311] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:32.068 [2024-09-29 16:52:53.739408] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:32.068 [2024-09-29 16:52:53.739458] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:32.327 pt3 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:32.327 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:32.328 "name": "raid_bdev1", 00:14:32.328 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:32.328 "strip_size_kb": 64, 00:14:32.328 "state": "configuring", 00:14:32.328 "raid_level": "raid5f", 00:14:32.328 "superblock": true, 00:14:32.328 "num_base_bdevs": 4, 00:14:32.328 "num_base_bdevs_discovered": 2, 00:14:32.328 "num_base_bdevs_operational": 3, 00:14:32.328 "base_bdevs_list": [ 00:14:32.328 { 00:14:32.328 "name": null, 00:14:32.328 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.328 "is_configured": false, 00:14:32.328 "data_offset": 2048, 00:14:32.328 "data_size": 63488 00:14:32.328 }, 00:14:32.328 { 00:14:32.328 "name": "pt2", 00:14:32.328 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:32.328 "is_configured": true, 00:14:32.328 "data_offset": 2048, 00:14:32.328 "data_size": 63488 00:14:32.328 }, 00:14:32.328 { 00:14:32.328 "name": "pt3", 00:14:32.328 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:32.328 "is_configured": true, 00:14:32.328 "data_offset": 2048, 00:14:32.328 "data_size": 63488 00:14:32.328 }, 00:14:32.328 { 00:14:32.328 "name": null, 00:14:32.328 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:32.328 "is_configured": false, 00:14:32.328 "data_offset": 2048, 00:14:32.328 "data_size": 63488 00:14:32.328 } 00:14:32.328 ] 00:14:32.328 }' 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:32.328 16:52:53 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.587 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:32.587 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:32.587 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:14:32.587 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:32.587 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.587 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.587 [2024-09-29 16:52:54.185893] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:32.587 [2024-09-29 16:52:54.185943] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:32.587 [2024-09-29 16:52:54.185961] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:32.587 [2024-09-29 16:52:54.185972] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:32.587 [2024-09-29 16:52:54.186303] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:32.587 [2024-09-29 16:52:54.186321] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:32.587 [2024-09-29 16:52:54.186379] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:32.587 [2024-09-29 16:52:54.186399] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:32.587 [2024-09-29 16:52:54.186485] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:32.587 [2024-09-29 16:52:54.186495] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:32.587 [2024-09-29 16:52:54.186707] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:14:32.587 [2024-09-29 16:52:54.187275] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:32.587 [2024-09-29 16:52:54.187327] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:14:32.588 [2024-09-29 16:52:54.187549] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:32.588 pt4 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:32.588 "name": "raid_bdev1", 00:14:32.588 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:32.588 "strip_size_kb": 64, 00:14:32.588 "state": "online", 00:14:32.588 "raid_level": "raid5f", 00:14:32.588 "superblock": true, 00:14:32.588 "num_base_bdevs": 4, 00:14:32.588 "num_base_bdevs_discovered": 3, 00:14:32.588 "num_base_bdevs_operational": 3, 00:14:32.588 "base_bdevs_list": [ 00:14:32.588 { 00:14:32.588 "name": null, 00:14:32.588 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.588 "is_configured": false, 00:14:32.588 "data_offset": 2048, 00:14:32.588 "data_size": 63488 00:14:32.588 }, 00:14:32.588 { 00:14:32.588 "name": "pt2", 00:14:32.588 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:32.588 "is_configured": true, 00:14:32.588 "data_offset": 2048, 00:14:32.588 "data_size": 63488 00:14:32.588 }, 00:14:32.588 { 00:14:32.588 "name": "pt3", 00:14:32.588 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:32.588 "is_configured": true, 00:14:32.588 "data_offset": 2048, 00:14:32.588 "data_size": 63488 00:14:32.588 }, 00:14:32.588 { 00:14:32.588 "name": "pt4", 00:14:32.588 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:32.588 "is_configured": true, 00:14:32.588 "data_offset": 2048, 00:14:32.588 "data_size": 63488 00:14:32.588 } 00:14:32.588 ] 00:14:32.588 }' 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:32.588 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.158 [2024-09-29 16:52:54.629114] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:33.158 [2024-09-29 16:52:54.629183] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:33.158 [2024-09-29 16:52:54.629256] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:33.158 [2024-09-29 16:52:54.629364] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:33.158 [2024-09-29 16:52:54.629418] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.158 [2024-09-29 16:52:54.685035] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:33.158 [2024-09-29 16:52:54.685128] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:33.158 [2024-09-29 16:52:54.685160] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:14:33.158 [2024-09-29 16:52:54.685186] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:33.158 [2024-09-29 16:52:54.687401] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:33.158 [2024-09-29 16:52:54.687486] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:33.158 [2024-09-29 16:52:54.687567] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:33.158 [2024-09-29 16:52:54.687623] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:33.158 [2024-09-29 16:52:54.687793] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:33.158 [2024-09-29 16:52:54.687850] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:33.158 [2024-09-29 16:52:54.687908] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:14:33.158 [2024-09-29 16:52:54.688020] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:33.158 [2024-09-29 16:52:54.688167] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:33.158 pt1 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.158 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.158 "name": "raid_bdev1", 00:14:33.158 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:33.158 "strip_size_kb": 64, 00:14:33.158 "state": "configuring", 00:14:33.158 "raid_level": "raid5f", 00:14:33.158 "superblock": true, 00:14:33.158 "num_base_bdevs": 4, 00:14:33.158 "num_base_bdevs_discovered": 2, 00:14:33.158 "num_base_bdevs_operational": 3, 00:14:33.158 "base_bdevs_list": [ 00:14:33.158 { 00:14:33.158 "name": null, 00:14:33.158 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.158 "is_configured": false, 00:14:33.158 "data_offset": 2048, 00:14:33.158 "data_size": 63488 00:14:33.158 }, 00:14:33.158 { 00:14:33.158 "name": "pt2", 00:14:33.158 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:33.158 "is_configured": true, 00:14:33.158 "data_offset": 2048, 00:14:33.158 "data_size": 63488 00:14:33.158 }, 00:14:33.158 { 00:14:33.158 "name": "pt3", 00:14:33.158 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:33.158 "is_configured": true, 00:14:33.158 "data_offset": 2048, 00:14:33.158 "data_size": 63488 00:14:33.158 }, 00:14:33.158 { 00:14:33.159 "name": null, 00:14:33.159 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:33.159 "is_configured": false, 00:14:33.159 "data_offset": 2048, 00:14:33.159 "data_size": 63488 00:14:33.159 } 00:14:33.159 ] 00:14:33.159 }' 00:14:33.159 16:52:54 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.159 16:52:54 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.729 [2024-09-29 16:52:55.180229] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:14:33.729 [2024-09-29 16:52:55.180320] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:33.729 [2024-09-29 16:52:55.180341] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:14:33.729 [2024-09-29 16:52:55.180352] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:33.729 [2024-09-29 16:52:55.180688] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:33.729 [2024-09-29 16:52:55.180709] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:14:33.729 [2024-09-29 16:52:55.180776] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:14:33.729 [2024-09-29 16:52:55.180806] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:14:33.729 [2024-09-29 16:52:55.180902] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:14:33.729 [2024-09-29 16:52:55.180913] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:33.729 [2024-09-29 16:52:55.181134] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:33.729 [2024-09-29 16:52:55.181670] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:14:33.729 [2024-09-29 16:52:55.181691] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:14:33.729 [2024-09-29 16:52:55.181876] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:33.729 pt4 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:33.729 "name": "raid_bdev1", 00:14:33.729 "uuid": "8255bd53-cbbc-4092-9fdd-c69fd5ade7cd", 00:14:33.729 "strip_size_kb": 64, 00:14:33.729 "state": "online", 00:14:33.729 "raid_level": "raid5f", 00:14:33.729 "superblock": true, 00:14:33.729 "num_base_bdevs": 4, 00:14:33.729 "num_base_bdevs_discovered": 3, 00:14:33.729 "num_base_bdevs_operational": 3, 00:14:33.729 "base_bdevs_list": [ 00:14:33.729 { 00:14:33.729 "name": null, 00:14:33.729 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:33.729 "is_configured": false, 00:14:33.729 "data_offset": 2048, 00:14:33.729 "data_size": 63488 00:14:33.729 }, 00:14:33.729 { 00:14:33.729 "name": "pt2", 00:14:33.729 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:33.729 "is_configured": true, 00:14:33.729 "data_offset": 2048, 00:14:33.729 "data_size": 63488 00:14:33.729 }, 00:14:33.729 { 00:14:33.729 "name": "pt3", 00:14:33.729 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:33.729 "is_configured": true, 00:14:33.729 "data_offset": 2048, 00:14:33.729 "data_size": 63488 00:14:33.729 }, 00:14:33.729 { 00:14:33.729 "name": "pt4", 00:14:33.729 "uuid": "00000000-0000-0000-0000-000000000004", 00:14:33.729 "is_configured": true, 00:14:33.729 "data_offset": 2048, 00:14:33.729 "data_size": 63488 00:14:33.729 } 00:14:33.729 ] 00:14:33.729 }' 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:33.729 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.989 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:33.989 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:33.989 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.989 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:33.989 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.249 [2024-09-29 16:52:55.683537] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 8255bd53-cbbc-4092-9fdd-c69fd5ade7cd '!=' 8255bd53-cbbc-4092-9fdd-c69fd5ade7cd ']' 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 94197 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 94197 ']' 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 94197 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 94197 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:34.249 killing process with pid 94197 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 94197' 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 94197 00:14:34.249 [2024-09-29 16:52:55.772255] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:34.249 [2024-09-29 16:52:55.772334] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:34.249 [2024-09-29 16:52:55.772402] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:34.249 [2024-09-29 16:52:55.772412] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:14:34.249 16:52:55 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 94197 00:14:34.249 [2024-09-29 16:52:55.815712] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:34.509 16:52:56 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:14:34.509 00:14:34.509 real 0m7.225s 00:14:34.509 user 0m12.123s 00:14:34.509 sys 0m1.637s 00:14:34.509 16:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:34.509 16:52:56 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.509 ************************************ 00:14:34.509 END TEST raid5f_superblock_test 00:14:34.509 ************************************ 00:14:34.509 16:52:56 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:14:34.509 16:52:56 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:14:34.509 16:52:56 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:14:34.509 16:52:56 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:34.509 16:52:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:34.509 ************************************ 00:14:34.509 START TEST raid5f_rebuild_test 00:14:34.509 ************************************ 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 false false true 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:34.509 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=94670 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 94670 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 94670 ']' 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:34.510 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:34.510 16:52:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:34.769 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:34.769 Zero copy mechanism will not be used. 00:14:34.769 [2024-09-29 16:52:56.256979] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:14:34.769 [2024-09-29 16:52:56.257121] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94670 ] 00:14:34.769 [2024-09-29 16:52:56.404697] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:35.029 [2024-09-29 16:52:56.451791] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:35.029 [2024-09-29 16:52:56.494414] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:35.029 [2024-09-29 16:52:56.494453] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.598 BaseBdev1_malloc 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.598 [2024-09-29 16:52:57.108702] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:35.598 [2024-09-29 16:52:57.108763] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.598 [2024-09-29 16:52:57.108784] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:35.598 [2024-09-29 16:52:57.108796] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.598 [2024-09-29 16:52:57.110890] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.598 [2024-09-29 16:52:57.110929] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:35.598 BaseBdev1 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.598 BaseBdev2_malloc 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.598 [2024-09-29 16:52:57.153206] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:35.598 [2024-09-29 16:52:57.153304] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.598 [2024-09-29 16:52:57.153348] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:35.598 [2024-09-29 16:52:57.153369] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.598 [2024-09-29 16:52:57.158229] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.598 [2024-09-29 16:52:57.158302] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:35.598 BaseBdev2 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.598 BaseBdev3_malloc 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.598 [2024-09-29 16:52:57.184454] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:35.598 [2024-09-29 16:52:57.184510] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.598 [2024-09-29 16:52:57.184533] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:35.598 [2024-09-29 16:52:57.184541] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.598 [2024-09-29 16:52:57.186597] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.598 [2024-09-29 16:52:57.186686] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:35.598 BaseBdev3 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.598 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.599 BaseBdev4_malloc 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.599 [2024-09-29 16:52:57.213271] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:35.599 [2024-09-29 16:52:57.213369] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.599 [2024-09-29 16:52:57.213396] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:35.599 [2024-09-29 16:52:57.213404] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.599 [2024-09-29 16:52:57.215460] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.599 [2024-09-29 16:52:57.215498] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:35.599 BaseBdev4 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.599 spare_malloc 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.599 spare_delay 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.599 [2024-09-29 16:52:57.253732] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:35.599 [2024-09-29 16:52:57.253774] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:35.599 [2024-09-29 16:52:57.253792] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:35.599 [2024-09-29 16:52:57.253800] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:35.599 [2024-09-29 16:52:57.255846] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:35.599 [2024-09-29 16:52:57.255930] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:35.599 spare 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.599 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.599 [2024-09-29 16:52:57.265796] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:35.599 [2024-09-29 16:52:57.267694] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:35.599 [2024-09-29 16:52:57.267778] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:35.599 [2024-09-29 16:52:57.267833] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:35.599 [2024-09-29 16:52:57.267920] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:35.599 [2024-09-29 16:52:57.267930] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:35.599 [2024-09-29 16:52:57.268221] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:35.599 [2024-09-29 16:52:57.268704] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:35.599 [2024-09-29 16:52:57.268752] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:35.599 [2024-09-29 16:52:57.268881] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.858 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:35.858 "name": "raid_bdev1", 00:14:35.858 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:35.858 "strip_size_kb": 64, 00:14:35.858 "state": "online", 00:14:35.858 "raid_level": "raid5f", 00:14:35.858 "superblock": false, 00:14:35.859 "num_base_bdevs": 4, 00:14:35.859 "num_base_bdevs_discovered": 4, 00:14:35.859 "num_base_bdevs_operational": 4, 00:14:35.859 "base_bdevs_list": [ 00:14:35.859 { 00:14:35.859 "name": "BaseBdev1", 00:14:35.859 "uuid": "086fa8a0-3eef-59be-a0d0-61e187549030", 00:14:35.859 "is_configured": true, 00:14:35.859 "data_offset": 0, 00:14:35.859 "data_size": 65536 00:14:35.859 }, 00:14:35.859 { 00:14:35.859 "name": "BaseBdev2", 00:14:35.859 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:35.859 "is_configured": true, 00:14:35.859 "data_offset": 0, 00:14:35.859 "data_size": 65536 00:14:35.859 }, 00:14:35.859 { 00:14:35.859 "name": "BaseBdev3", 00:14:35.859 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:35.859 "is_configured": true, 00:14:35.859 "data_offset": 0, 00:14:35.859 "data_size": 65536 00:14:35.859 }, 00:14:35.859 { 00:14:35.859 "name": "BaseBdev4", 00:14:35.859 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:35.859 "is_configured": true, 00:14:35.859 "data_offset": 0, 00:14:35.859 "data_size": 65536 00:14:35.859 } 00:14:35.859 ] 00:14:35.859 }' 00:14:35.859 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:35.859 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.118 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:36.118 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:36.118 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:36.118 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.118 [2024-09-29 16:52:57.738018] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:36.118 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:36.118 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:14:36.119 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.119 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:36.119 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:36.119 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:36.119 16:52:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:36.379 16:52:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:36.379 [2024-09-29 16:52:58.009416] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:36.379 /dev/nbd0 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:36.639 1+0 records in 00:14:36.639 1+0 records out 00:14:36.639 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000637839 s, 6.4 MB/s 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:36.639 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:14:37.209 512+0 records in 00:14:37.209 512+0 records out 00:14:37.209 100663296 bytes (101 MB, 96 MiB) copied, 0.555274 s, 181 MB/s 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:37.209 [2024-09-29 16:52:58.864861] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.209 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.469 [2024-09-29 16:52:58.885040] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:37.469 "name": "raid_bdev1", 00:14:37.469 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:37.469 "strip_size_kb": 64, 00:14:37.469 "state": "online", 00:14:37.469 "raid_level": "raid5f", 00:14:37.469 "superblock": false, 00:14:37.469 "num_base_bdevs": 4, 00:14:37.469 "num_base_bdevs_discovered": 3, 00:14:37.469 "num_base_bdevs_operational": 3, 00:14:37.469 "base_bdevs_list": [ 00:14:37.469 { 00:14:37.469 "name": null, 00:14:37.469 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:37.469 "is_configured": false, 00:14:37.469 "data_offset": 0, 00:14:37.469 "data_size": 65536 00:14:37.469 }, 00:14:37.469 { 00:14:37.469 "name": "BaseBdev2", 00:14:37.469 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:37.469 "is_configured": true, 00:14:37.469 "data_offset": 0, 00:14:37.469 "data_size": 65536 00:14:37.469 }, 00:14:37.469 { 00:14:37.469 "name": "BaseBdev3", 00:14:37.469 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:37.469 "is_configured": true, 00:14:37.469 "data_offset": 0, 00:14:37.469 "data_size": 65536 00:14:37.469 }, 00:14:37.469 { 00:14:37.469 "name": "BaseBdev4", 00:14:37.469 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:37.469 "is_configured": true, 00:14:37.469 "data_offset": 0, 00:14:37.469 "data_size": 65536 00:14:37.469 } 00:14:37.469 ] 00:14:37.469 }' 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:37.469 16:52:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.727 16:52:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:37.727 16:52:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.727 16:52:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:37.727 [2024-09-29 16:52:59.340283] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:37.727 [2024-09-29 16:52:59.343763] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:14:37.727 [2024-09-29 16:52:59.345972] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:37.727 16:52:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.727 16:52:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:39.106 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:39.106 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.106 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:39.106 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:39.106 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:39.107 "name": "raid_bdev1", 00:14:39.107 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:39.107 "strip_size_kb": 64, 00:14:39.107 "state": "online", 00:14:39.107 "raid_level": "raid5f", 00:14:39.107 "superblock": false, 00:14:39.107 "num_base_bdevs": 4, 00:14:39.107 "num_base_bdevs_discovered": 4, 00:14:39.107 "num_base_bdevs_operational": 4, 00:14:39.107 "process": { 00:14:39.107 "type": "rebuild", 00:14:39.107 "target": "spare", 00:14:39.107 "progress": { 00:14:39.107 "blocks": 19200, 00:14:39.107 "percent": 9 00:14:39.107 } 00:14:39.107 }, 00:14:39.107 "base_bdevs_list": [ 00:14:39.107 { 00:14:39.107 "name": "spare", 00:14:39.107 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:39.107 "is_configured": true, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 }, 00:14:39.107 { 00:14:39.107 "name": "BaseBdev2", 00:14:39.107 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:39.107 "is_configured": true, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 }, 00:14:39.107 { 00:14:39.107 "name": "BaseBdev3", 00:14:39.107 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:39.107 "is_configured": true, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 }, 00:14:39.107 { 00:14:39.107 "name": "BaseBdev4", 00:14:39.107 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:39.107 "is_configured": true, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 } 00:14:39.107 ] 00:14:39.107 }' 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.107 [2024-09-29 16:53:00.512775] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:39.107 [2024-09-29 16:53:00.551544] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:39.107 [2024-09-29 16:53:00.551642] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:39.107 [2024-09-29 16:53:00.551663] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:39.107 [2024-09-29 16:53:00.551672] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.107 "name": "raid_bdev1", 00:14:39.107 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:39.107 "strip_size_kb": 64, 00:14:39.107 "state": "online", 00:14:39.107 "raid_level": "raid5f", 00:14:39.107 "superblock": false, 00:14:39.107 "num_base_bdevs": 4, 00:14:39.107 "num_base_bdevs_discovered": 3, 00:14:39.107 "num_base_bdevs_operational": 3, 00:14:39.107 "base_bdevs_list": [ 00:14:39.107 { 00:14:39.107 "name": null, 00:14:39.107 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.107 "is_configured": false, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 }, 00:14:39.107 { 00:14:39.107 "name": "BaseBdev2", 00:14:39.107 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:39.107 "is_configured": true, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 }, 00:14:39.107 { 00:14:39.107 "name": "BaseBdev3", 00:14:39.107 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:39.107 "is_configured": true, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 }, 00:14:39.107 { 00:14:39.107 "name": "BaseBdev4", 00:14:39.107 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:39.107 "is_configured": true, 00:14:39.107 "data_offset": 0, 00:14:39.107 "data_size": 65536 00:14:39.107 } 00:14:39.107 ] 00:14:39.107 }' 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.107 16:53:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:39.367 "name": "raid_bdev1", 00:14:39.367 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:39.367 "strip_size_kb": 64, 00:14:39.367 "state": "online", 00:14:39.367 "raid_level": "raid5f", 00:14:39.367 "superblock": false, 00:14:39.367 "num_base_bdevs": 4, 00:14:39.367 "num_base_bdevs_discovered": 3, 00:14:39.367 "num_base_bdevs_operational": 3, 00:14:39.367 "base_bdevs_list": [ 00:14:39.367 { 00:14:39.367 "name": null, 00:14:39.367 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:39.367 "is_configured": false, 00:14:39.367 "data_offset": 0, 00:14:39.367 "data_size": 65536 00:14:39.367 }, 00:14:39.367 { 00:14:39.367 "name": "BaseBdev2", 00:14:39.367 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:39.367 "is_configured": true, 00:14:39.367 "data_offset": 0, 00:14:39.367 "data_size": 65536 00:14:39.367 }, 00:14:39.367 { 00:14:39.367 "name": "BaseBdev3", 00:14:39.367 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:39.367 "is_configured": true, 00:14:39.367 "data_offset": 0, 00:14:39.367 "data_size": 65536 00:14:39.367 }, 00:14:39.367 { 00:14:39.367 "name": "BaseBdev4", 00:14:39.367 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:39.367 "is_configured": true, 00:14:39.367 "data_offset": 0, 00:14:39.367 "data_size": 65536 00:14:39.367 } 00:14:39.367 ] 00:14:39.367 }' 00:14:39.367 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:39.627 [2024-09-29 16:53:01.139966] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:39.627 [2024-09-29 16:53:01.143097] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027e70 00:14:39.627 [2024-09-29 16:53:01.145169] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.627 16:53:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:40.567 "name": "raid_bdev1", 00:14:40.567 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:40.567 "strip_size_kb": 64, 00:14:40.567 "state": "online", 00:14:40.567 "raid_level": "raid5f", 00:14:40.567 "superblock": false, 00:14:40.567 "num_base_bdevs": 4, 00:14:40.567 "num_base_bdevs_discovered": 4, 00:14:40.567 "num_base_bdevs_operational": 4, 00:14:40.567 "process": { 00:14:40.567 "type": "rebuild", 00:14:40.567 "target": "spare", 00:14:40.567 "progress": { 00:14:40.567 "blocks": 19200, 00:14:40.567 "percent": 9 00:14:40.567 } 00:14:40.567 }, 00:14:40.567 "base_bdevs_list": [ 00:14:40.567 { 00:14:40.567 "name": "spare", 00:14:40.567 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:40.567 "is_configured": true, 00:14:40.567 "data_offset": 0, 00:14:40.567 "data_size": 65536 00:14:40.567 }, 00:14:40.567 { 00:14:40.567 "name": "BaseBdev2", 00:14:40.567 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:40.567 "is_configured": true, 00:14:40.567 "data_offset": 0, 00:14:40.567 "data_size": 65536 00:14:40.567 }, 00:14:40.567 { 00:14:40.567 "name": "BaseBdev3", 00:14:40.567 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:40.567 "is_configured": true, 00:14:40.567 "data_offset": 0, 00:14:40.567 "data_size": 65536 00:14:40.567 }, 00:14:40.567 { 00:14:40.567 "name": "BaseBdev4", 00:14:40.567 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:40.567 "is_configured": true, 00:14:40.567 "data_offset": 0, 00:14:40.567 "data_size": 65536 00:14:40.567 } 00:14:40.567 ] 00:14:40.567 }' 00:14:40.567 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=507 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:40.827 "name": "raid_bdev1", 00:14:40.827 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:40.827 "strip_size_kb": 64, 00:14:40.827 "state": "online", 00:14:40.827 "raid_level": "raid5f", 00:14:40.827 "superblock": false, 00:14:40.827 "num_base_bdevs": 4, 00:14:40.827 "num_base_bdevs_discovered": 4, 00:14:40.827 "num_base_bdevs_operational": 4, 00:14:40.827 "process": { 00:14:40.827 "type": "rebuild", 00:14:40.827 "target": "spare", 00:14:40.827 "progress": { 00:14:40.827 "blocks": 21120, 00:14:40.827 "percent": 10 00:14:40.827 } 00:14:40.827 }, 00:14:40.827 "base_bdevs_list": [ 00:14:40.827 { 00:14:40.827 "name": "spare", 00:14:40.827 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:40.827 "is_configured": true, 00:14:40.827 "data_offset": 0, 00:14:40.827 "data_size": 65536 00:14:40.827 }, 00:14:40.827 { 00:14:40.827 "name": "BaseBdev2", 00:14:40.827 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:40.827 "is_configured": true, 00:14:40.827 "data_offset": 0, 00:14:40.827 "data_size": 65536 00:14:40.827 }, 00:14:40.827 { 00:14:40.827 "name": "BaseBdev3", 00:14:40.827 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:40.827 "is_configured": true, 00:14:40.827 "data_offset": 0, 00:14:40.827 "data_size": 65536 00:14:40.827 }, 00:14:40.827 { 00:14:40.827 "name": "BaseBdev4", 00:14:40.827 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:40.827 "is_configured": true, 00:14:40.827 "data_offset": 0, 00:14:40.827 "data_size": 65536 00:14:40.827 } 00:14:40.827 ] 00:14:40.827 }' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:40.827 16:53:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:42.207 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:42.207 "name": "raid_bdev1", 00:14:42.207 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:42.207 "strip_size_kb": 64, 00:14:42.207 "state": "online", 00:14:42.207 "raid_level": "raid5f", 00:14:42.207 "superblock": false, 00:14:42.207 "num_base_bdevs": 4, 00:14:42.207 "num_base_bdevs_discovered": 4, 00:14:42.207 "num_base_bdevs_operational": 4, 00:14:42.207 "process": { 00:14:42.207 "type": "rebuild", 00:14:42.207 "target": "spare", 00:14:42.208 "progress": { 00:14:42.208 "blocks": 44160, 00:14:42.208 "percent": 22 00:14:42.208 } 00:14:42.208 }, 00:14:42.208 "base_bdevs_list": [ 00:14:42.208 { 00:14:42.208 "name": "spare", 00:14:42.208 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:42.208 "is_configured": true, 00:14:42.208 "data_offset": 0, 00:14:42.208 "data_size": 65536 00:14:42.208 }, 00:14:42.208 { 00:14:42.208 "name": "BaseBdev2", 00:14:42.208 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:42.208 "is_configured": true, 00:14:42.208 "data_offset": 0, 00:14:42.208 "data_size": 65536 00:14:42.208 }, 00:14:42.208 { 00:14:42.208 "name": "BaseBdev3", 00:14:42.208 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:42.208 "is_configured": true, 00:14:42.208 "data_offset": 0, 00:14:42.208 "data_size": 65536 00:14:42.208 }, 00:14:42.208 { 00:14:42.208 "name": "BaseBdev4", 00:14:42.208 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:42.208 "is_configured": true, 00:14:42.208 "data_offset": 0, 00:14:42.208 "data_size": 65536 00:14:42.208 } 00:14:42.208 ] 00:14:42.208 }' 00:14:42.208 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:42.208 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:42.208 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:42.208 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:42.208 16:53:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:43.145 16:53:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:43.146 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:43.146 "name": "raid_bdev1", 00:14:43.146 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:43.146 "strip_size_kb": 64, 00:14:43.146 "state": "online", 00:14:43.146 "raid_level": "raid5f", 00:14:43.146 "superblock": false, 00:14:43.146 "num_base_bdevs": 4, 00:14:43.146 "num_base_bdevs_discovered": 4, 00:14:43.146 "num_base_bdevs_operational": 4, 00:14:43.146 "process": { 00:14:43.146 "type": "rebuild", 00:14:43.146 "target": "spare", 00:14:43.146 "progress": { 00:14:43.146 "blocks": 65280, 00:14:43.146 "percent": 33 00:14:43.146 } 00:14:43.146 }, 00:14:43.146 "base_bdevs_list": [ 00:14:43.146 { 00:14:43.146 "name": "spare", 00:14:43.146 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:43.146 "is_configured": true, 00:14:43.146 "data_offset": 0, 00:14:43.146 "data_size": 65536 00:14:43.146 }, 00:14:43.146 { 00:14:43.146 "name": "BaseBdev2", 00:14:43.146 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:43.146 "is_configured": true, 00:14:43.146 "data_offset": 0, 00:14:43.146 "data_size": 65536 00:14:43.146 }, 00:14:43.146 { 00:14:43.146 "name": "BaseBdev3", 00:14:43.146 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:43.146 "is_configured": true, 00:14:43.146 "data_offset": 0, 00:14:43.146 "data_size": 65536 00:14:43.146 }, 00:14:43.146 { 00:14:43.146 "name": "BaseBdev4", 00:14:43.146 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:43.146 "is_configured": true, 00:14:43.146 "data_offset": 0, 00:14:43.146 "data_size": 65536 00:14:43.146 } 00:14:43.146 ] 00:14:43.146 }' 00:14:43.146 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:43.146 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:43.146 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:43.146 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:43.146 16:53:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:44.527 "name": "raid_bdev1", 00:14:44.527 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:44.527 "strip_size_kb": 64, 00:14:44.527 "state": "online", 00:14:44.527 "raid_level": "raid5f", 00:14:44.527 "superblock": false, 00:14:44.527 "num_base_bdevs": 4, 00:14:44.527 "num_base_bdevs_discovered": 4, 00:14:44.527 "num_base_bdevs_operational": 4, 00:14:44.527 "process": { 00:14:44.527 "type": "rebuild", 00:14:44.527 "target": "spare", 00:14:44.527 "progress": { 00:14:44.527 "blocks": 88320, 00:14:44.527 "percent": 44 00:14:44.527 } 00:14:44.527 }, 00:14:44.527 "base_bdevs_list": [ 00:14:44.527 { 00:14:44.527 "name": "spare", 00:14:44.527 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:44.527 "is_configured": true, 00:14:44.527 "data_offset": 0, 00:14:44.527 "data_size": 65536 00:14:44.527 }, 00:14:44.527 { 00:14:44.527 "name": "BaseBdev2", 00:14:44.527 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:44.527 "is_configured": true, 00:14:44.527 "data_offset": 0, 00:14:44.527 "data_size": 65536 00:14:44.527 }, 00:14:44.527 { 00:14:44.527 "name": "BaseBdev3", 00:14:44.527 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:44.527 "is_configured": true, 00:14:44.527 "data_offset": 0, 00:14:44.527 "data_size": 65536 00:14:44.527 }, 00:14:44.527 { 00:14:44.527 "name": "BaseBdev4", 00:14:44.527 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:44.527 "is_configured": true, 00:14:44.527 "data_offset": 0, 00:14:44.527 "data_size": 65536 00:14:44.527 } 00:14:44.527 ] 00:14:44.527 }' 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:44.527 16:53:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.467 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:45.467 "name": "raid_bdev1", 00:14:45.467 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:45.467 "strip_size_kb": 64, 00:14:45.467 "state": "online", 00:14:45.467 "raid_level": "raid5f", 00:14:45.467 "superblock": false, 00:14:45.467 "num_base_bdevs": 4, 00:14:45.467 "num_base_bdevs_discovered": 4, 00:14:45.467 "num_base_bdevs_operational": 4, 00:14:45.467 "process": { 00:14:45.467 "type": "rebuild", 00:14:45.467 "target": "spare", 00:14:45.467 "progress": { 00:14:45.467 "blocks": 109440, 00:14:45.467 "percent": 55 00:14:45.467 } 00:14:45.467 }, 00:14:45.467 "base_bdevs_list": [ 00:14:45.467 { 00:14:45.467 "name": "spare", 00:14:45.467 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:45.467 "is_configured": true, 00:14:45.467 "data_offset": 0, 00:14:45.467 "data_size": 65536 00:14:45.467 }, 00:14:45.467 { 00:14:45.467 "name": "BaseBdev2", 00:14:45.467 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:45.467 "is_configured": true, 00:14:45.467 "data_offset": 0, 00:14:45.467 "data_size": 65536 00:14:45.467 }, 00:14:45.468 { 00:14:45.468 "name": "BaseBdev3", 00:14:45.468 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:45.468 "is_configured": true, 00:14:45.468 "data_offset": 0, 00:14:45.468 "data_size": 65536 00:14:45.468 }, 00:14:45.468 { 00:14:45.468 "name": "BaseBdev4", 00:14:45.468 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:45.468 "is_configured": true, 00:14:45.468 "data_offset": 0, 00:14:45.468 "data_size": 65536 00:14:45.468 } 00:14:45.468 ] 00:14:45.468 }' 00:14:45.468 16:53:06 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:45.468 16:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:45.468 16:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:45.468 16:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:45.468 16:53:07 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:46.848 "name": "raid_bdev1", 00:14:46.848 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:46.848 "strip_size_kb": 64, 00:14:46.848 "state": "online", 00:14:46.848 "raid_level": "raid5f", 00:14:46.848 "superblock": false, 00:14:46.848 "num_base_bdevs": 4, 00:14:46.848 "num_base_bdevs_discovered": 4, 00:14:46.848 "num_base_bdevs_operational": 4, 00:14:46.848 "process": { 00:14:46.848 "type": "rebuild", 00:14:46.848 "target": "spare", 00:14:46.848 "progress": { 00:14:46.848 "blocks": 132480, 00:14:46.848 "percent": 67 00:14:46.848 } 00:14:46.848 }, 00:14:46.848 "base_bdevs_list": [ 00:14:46.848 { 00:14:46.848 "name": "spare", 00:14:46.848 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:46.848 "is_configured": true, 00:14:46.848 "data_offset": 0, 00:14:46.848 "data_size": 65536 00:14:46.848 }, 00:14:46.848 { 00:14:46.848 "name": "BaseBdev2", 00:14:46.848 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:46.848 "is_configured": true, 00:14:46.848 "data_offset": 0, 00:14:46.848 "data_size": 65536 00:14:46.848 }, 00:14:46.848 { 00:14:46.848 "name": "BaseBdev3", 00:14:46.848 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:46.848 "is_configured": true, 00:14:46.848 "data_offset": 0, 00:14:46.848 "data_size": 65536 00:14:46.848 }, 00:14:46.848 { 00:14:46.848 "name": "BaseBdev4", 00:14:46.848 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:46.848 "is_configured": true, 00:14:46.848 "data_offset": 0, 00:14:46.848 "data_size": 65536 00:14:46.848 } 00:14:46.848 ] 00:14:46.848 }' 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:46.848 16:53:08 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.787 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:47.787 "name": "raid_bdev1", 00:14:47.787 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:47.787 "strip_size_kb": 64, 00:14:47.787 "state": "online", 00:14:47.787 "raid_level": "raid5f", 00:14:47.787 "superblock": false, 00:14:47.787 "num_base_bdevs": 4, 00:14:47.787 "num_base_bdevs_discovered": 4, 00:14:47.787 "num_base_bdevs_operational": 4, 00:14:47.787 "process": { 00:14:47.787 "type": "rebuild", 00:14:47.787 "target": "spare", 00:14:47.787 "progress": { 00:14:47.787 "blocks": 153600, 00:14:47.787 "percent": 78 00:14:47.787 } 00:14:47.788 }, 00:14:47.788 "base_bdevs_list": [ 00:14:47.788 { 00:14:47.788 "name": "spare", 00:14:47.788 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:47.788 "is_configured": true, 00:14:47.788 "data_offset": 0, 00:14:47.788 "data_size": 65536 00:14:47.788 }, 00:14:47.788 { 00:14:47.788 "name": "BaseBdev2", 00:14:47.788 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:47.788 "is_configured": true, 00:14:47.788 "data_offset": 0, 00:14:47.788 "data_size": 65536 00:14:47.788 }, 00:14:47.788 { 00:14:47.788 "name": "BaseBdev3", 00:14:47.788 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:47.788 "is_configured": true, 00:14:47.788 "data_offset": 0, 00:14:47.788 "data_size": 65536 00:14:47.788 }, 00:14:47.788 { 00:14:47.788 "name": "BaseBdev4", 00:14:47.788 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:47.788 "is_configured": true, 00:14:47.788 "data_offset": 0, 00:14:47.788 "data_size": 65536 00:14:47.788 } 00:14:47.788 ] 00:14:47.788 }' 00:14:47.788 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:47.788 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:47.788 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:47.788 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:47.788 16:53:09 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:49.169 "name": "raid_bdev1", 00:14:49.169 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:49.169 "strip_size_kb": 64, 00:14:49.169 "state": "online", 00:14:49.169 "raid_level": "raid5f", 00:14:49.169 "superblock": false, 00:14:49.169 "num_base_bdevs": 4, 00:14:49.169 "num_base_bdevs_discovered": 4, 00:14:49.169 "num_base_bdevs_operational": 4, 00:14:49.169 "process": { 00:14:49.169 "type": "rebuild", 00:14:49.169 "target": "spare", 00:14:49.169 "progress": { 00:14:49.169 "blocks": 176640, 00:14:49.169 "percent": 89 00:14:49.169 } 00:14:49.169 }, 00:14:49.169 "base_bdevs_list": [ 00:14:49.169 { 00:14:49.169 "name": "spare", 00:14:49.169 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:49.169 "is_configured": true, 00:14:49.169 "data_offset": 0, 00:14:49.169 "data_size": 65536 00:14:49.169 }, 00:14:49.169 { 00:14:49.169 "name": "BaseBdev2", 00:14:49.169 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:49.169 "is_configured": true, 00:14:49.169 "data_offset": 0, 00:14:49.169 "data_size": 65536 00:14:49.169 }, 00:14:49.169 { 00:14:49.169 "name": "BaseBdev3", 00:14:49.169 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:49.169 "is_configured": true, 00:14:49.169 "data_offset": 0, 00:14:49.169 "data_size": 65536 00:14:49.169 }, 00:14:49.169 { 00:14:49.169 "name": "BaseBdev4", 00:14:49.169 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:49.169 "is_configured": true, 00:14:49.169 "data_offset": 0, 00:14:49.169 "data_size": 65536 00:14:49.169 } 00:14:49.169 ] 00:14:49.169 }' 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:49.169 16:53:10 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:50.111 [2024-09-29 16:53:11.485761] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:50.111 [2024-09-29 16:53:11.485875] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:50.111 [2024-09-29 16:53:11.485914] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:50.111 "name": "raid_bdev1", 00:14:50.111 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:50.111 "strip_size_kb": 64, 00:14:50.111 "state": "online", 00:14:50.111 "raid_level": "raid5f", 00:14:50.111 "superblock": false, 00:14:50.111 "num_base_bdevs": 4, 00:14:50.111 "num_base_bdevs_discovered": 4, 00:14:50.111 "num_base_bdevs_operational": 4, 00:14:50.111 "base_bdevs_list": [ 00:14:50.111 { 00:14:50.111 "name": "spare", 00:14:50.111 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 }, 00:14:50.111 { 00:14:50.111 "name": "BaseBdev2", 00:14:50.111 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 }, 00:14:50.111 { 00:14:50.111 "name": "BaseBdev3", 00:14:50.111 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 }, 00:14:50.111 { 00:14:50.111 "name": "BaseBdev4", 00:14:50.111 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 } 00:14:50.111 ] 00:14:50.111 }' 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:50.111 "name": "raid_bdev1", 00:14:50.111 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:50.111 "strip_size_kb": 64, 00:14:50.111 "state": "online", 00:14:50.111 "raid_level": "raid5f", 00:14:50.111 "superblock": false, 00:14:50.111 "num_base_bdevs": 4, 00:14:50.111 "num_base_bdevs_discovered": 4, 00:14:50.111 "num_base_bdevs_operational": 4, 00:14:50.111 "base_bdevs_list": [ 00:14:50.111 { 00:14:50.111 "name": "spare", 00:14:50.111 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 }, 00:14:50.111 { 00:14:50.111 "name": "BaseBdev2", 00:14:50.111 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 }, 00:14:50.111 { 00:14:50.111 "name": "BaseBdev3", 00:14:50.111 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 }, 00:14:50.111 { 00:14:50.111 "name": "BaseBdev4", 00:14:50.111 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:50.111 "is_configured": true, 00:14:50.111 "data_offset": 0, 00:14:50.111 "data_size": 65536 00:14:50.111 } 00:14:50.111 ] 00:14:50.111 }' 00:14:50.111 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.371 "name": "raid_bdev1", 00:14:50.371 "uuid": "6b7cec1c-f170-4838-aebe-504a03a1cde4", 00:14:50.371 "strip_size_kb": 64, 00:14:50.371 "state": "online", 00:14:50.371 "raid_level": "raid5f", 00:14:50.371 "superblock": false, 00:14:50.371 "num_base_bdevs": 4, 00:14:50.371 "num_base_bdevs_discovered": 4, 00:14:50.371 "num_base_bdevs_operational": 4, 00:14:50.371 "base_bdevs_list": [ 00:14:50.371 { 00:14:50.371 "name": "spare", 00:14:50.371 "uuid": "b8ddff53-ec00-5645-b395-85723ecc7f68", 00:14:50.371 "is_configured": true, 00:14:50.371 "data_offset": 0, 00:14:50.371 "data_size": 65536 00:14:50.371 }, 00:14:50.371 { 00:14:50.371 "name": "BaseBdev2", 00:14:50.371 "uuid": "c884131b-6c29-5f6c-8a33-5f2787446c33", 00:14:50.371 "is_configured": true, 00:14:50.371 "data_offset": 0, 00:14:50.371 "data_size": 65536 00:14:50.371 }, 00:14:50.371 { 00:14:50.371 "name": "BaseBdev3", 00:14:50.371 "uuid": "a2ec4115-7ca4-504a-b7d2-e533bf0efc4b", 00:14:50.371 "is_configured": true, 00:14:50.371 "data_offset": 0, 00:14:50.371 "data_size": 65536 00:14:50.371 }, 00:14:50.371 { 00:14:50.371 "name": "BaseBdev4", 00:14:50.371 "uuid": "0dff50a1-a5ea-52a7-8aa8-0727a7abfb0f", 00:14:50.371 "is_configured": true, 00:14:50.371 "data_offset": 0, 00:14:50.371 "data_size": 65536 00:14:50.371 } 00:14:50.371 ] 00:14:50.371 }' 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.371 16:53:11 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.631 [2024-09-29 16:53:12.269687] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:50.631 [2024-09-29 16:53:12.269716] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:50.631 [2024-09-29 16:53:12.269818] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:50.631 [2024-09-29 16:53:12.269912] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:50.631 [2024-09-29 16:53:12.269924] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.631 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:50.891 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:50.891 /dev/nbd0 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:51.151 1+0 records in 00:14:51.151 1+0 records out 00:14:51.151 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00025634 s, 16.0 MB/s 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:51.151 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:51.152 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:51.152 /dev/nbd1 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:51.411 1+0 records in 00:14:51.411 1+0 records out 00:14:51.411 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000208561 s, 19.6 MB/s 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:51.411 16:53:12 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:51.671 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 94670 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 94670 ']' 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 94670 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 94670 00:14:51.930 killing process with pid 94670 00:14:51.930 Received shutdown signal, test time was about 60.000000 seconds 00:14:51.930 00:14:51.930 Latency(us) 00:14:51.930 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:51.930 =================================================================================================================== 00:14:51.930 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 94670' 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 94670 00:14:51.930 [2024-09-29 16:53:13.401583] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:51.930 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 94670 00:14:51.930 [2024-09-29 16:53:13.451171] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:52.190 00:14:52.190 real 0m17.527s 00:14:52.190 user 0m21.280s 00:14:52.190 sys 0m2.459s 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:52.190 ************************************ 00:14:52.190 END TEST raid5f_rebuild_test 00:14:52.190 ************************************ 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.190 16:53:13 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:14:52.190 16:53:13 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:14:52.190 16:53:13 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:52.190 16:53:13 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:52.190 ************************************ 00:14:52.190 START TEST raid5f_rebuild_test_sb 00:14:52.190 ************************************ 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 true false true 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=95152 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 95152 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 95152 ']' 00:14:52.190 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:52.190 16:53:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:52.190 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:52.190 Zero copy mechanism will not be used. 00:14:52.191 [2024-09-29 16:53:13.858761] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:14:52.191 [2024-09-29 16:53:13.858881] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95152 ] 00:14:52.450 [2024-09-29 16:53:14.005529] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:52.450 [2024-09-29 16:53:14.050305] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:52.450 [2024-09-29 16:53:14.093127] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:52.450 [2024-09-29 16:53:14.093162] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.389 BaseBdev1_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.389 [2024-09-29 16:53:14.715249] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:53.389 [2024-09-29 16:53:14.715401] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:53.389 [2024-09-29 16:53:14.715434] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:53.389 [2024-09-29 16:53:14.715448] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:53.389 [2024-09-29 16:53:14.717539] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:53.389 [2024-09-29 16:53:14.717583] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:53.389 BaseBdev1 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.389 BaseBdev2_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.389 [2024-09-29 16:53:14.759864] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:53.389 [2024-09-29 16:53:14.759968] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:53.389 [2024-09-29 16:53:14.760012] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:53.389 [2024-09-29 16:53:14.760033] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:53.389 [2024-09-29 16:53:14.764877] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:53.389 [2024-09-29 16:53:14.765038] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:53.389 BaseBdev2 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.389 BaseBdev3_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.389 [2024-09-29 16:53:14.791344] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:53.389 [2024-09-29 16:53:14.791406] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:53.389 [2024-09-29 16:53:14.791446] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:53.389 [2024-09-29 16:53:14.791455] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:53.389 [2024-09-29 16:53:14.793534] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:53.389 [2024-09-29 16:53:14.793582] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:53.389 BaseBdev3 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.389 BaseBdev4_malloc 00:14:53.389 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.390 [2024-09-29 16:53:14.820160] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:14:53.390 [2024-09-29 16:53:14.820211] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:53.390 [2024-09-29 16:53:14.820234] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:53.390 [2024-09-29 16:53:14.820242] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:53.390 [2024-09-29 16:53:14.822256] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:53.390 [2024-09-29 16:53:14.822291] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:14:53.390 BaseBdev4 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.390 spare_malloc 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.390 spare_delay 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.390 [2024-09-29 16:53:14.860618] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:53.390 [2024-09-29 16:53:14.860666] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:53.390 [2024-09-29 16:53:14.860699] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:14:53.390 [2024-09-29 16:53:14.860708] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:53.390 [2024-09-29 16:53:14.862725] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:53.390 [2024-09-29 16:53:14.862777] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:53.390 spare 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.390 [2024-09-29 16:53:14.872682] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:53.390 [2024-09-29 16:53:14.874413] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:53.390 [2024-09-29 16:53:14.874544] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:53.390 [2024-09-29 16:53:14.874600] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:53.390 [2024-09-29 16:53:14.874789] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:53.390 [2024-09-29 16:53:14.874803] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:14:53.390 [2024-09-29 16:53:14.875044] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:53.390 [2024-09-29 16:53:14.875498] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:53.390 [2024-09-29 16:53:14.875512] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:53.390 [2024-09-29 16:53:14.875623] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.390 "name": "raid_bdev1", 00:14:53.390 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:53.390 "strip_size_kb": 64, 00:14:53.390 "state": "online", 00:14:53.390 "raid_level": "raid5f", 00:14:53.390 "superblock": true, 00:14:53.390 "num_base_bdevs": 4, 00:14:53.390 "num_base_bdevs_discovered": 4, 00:14:53.390 "num_base_bdevs_operational": 4, 00:14:53.390 "base_bdevs_list": [ 00:14:53.390 { 00:14:53.390 "name": "BaseBdev1", 00:14:53.390 "uuid": "43bc05aa-4db2-5fad-9385-3a2e11637c5c", 00:14:53.390 "is_configured": true, 00:14:53.390 "data_offset": 2048, 00:14:53.390 "data_size": 63488 00:14:53.390 }, 00:14:53.390 { 00:14:53.390 "name": "BaseBdev2", 00:14:53.390 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:53.390 "is_configured": true, 00:14:53.390 "data_offset": 2048, 00:14:53.390 "data_size": 63488 00:14:53.390 }, 00:14:53.390 { 00:14:53.390 "name": "BaseBdev3", 00:14:53.390 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:53.390 "is_configured": true, 00:14:53.390 "data_offset": 2048, 00:14:53.390 "data_size": 63488 00:14:53.390 }, 00:14:53.390 { 00:14:53.390 "name": "BaseBdev4", 00:14:53.390 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:53.390 "is_configured": true, 00:14:53.390 "data_offset": 2048, 00:14:53.390 "data_size": 63488 00:14:53.390 } 00:14:53.390 ] 00:14:53.390 }' 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.390 16:53:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.960 [2024-09-29 16:53:15.388662] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:53.960 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:54.225 [2024-09-29 16:53:15.640121] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:14:54.225 /dev/nbd0 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:54.225 1+0 records in 00:14:54.225 1+0 records out 00:14:54.225 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000493954 s, 8.3 MB/s 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:14:54.225 16:53:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:14:54.493 496+0 records in 00:14:54.493 496+0 records out 00:14:54.493 97517568 bytes (98 MB, 93 MiB) copied, 0.392746 s, 248 MB/s 00:14:54.493 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:54.493 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:54.493 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:54.493 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:54.493 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:54.493 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:54.493 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:54.754 [2024-09-29 16:53:16.296290] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.754 [2024-09-29 16:53:16.348249] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.754 "name": "raid_bdev1", 00:14:54.754 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:54.754 "strip_size_kb": 64, 00:14:54.754 "state": "online", 00:14:54.754 "raid_level": "raid5f", 00:14:54.754 "superblock": true, 00:14:54.754 "num_base_bdevs": 4, 00:14:54.754 "num_base_bdevs_discovered": 3, 00:14:54.754 "num_base_bdevs_operational": 3, 00:14:54.754 "base_bdevs_list": [ 00:14:54.754 { 00:14:54.754 "name": null, 00:14:54.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.754 "is_configured": false, 00:14:54.754 "data_offset": 0, 00:14:54.754 "data_size": 63488 00:14:54.754 }, 00:14:54.754 { 00:14:54.754 "name": "BaseBdev2", 00:14:54.754 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:54.754 "is_configured": true, 00:14:54.754 "data_offset": 2048, 00:14:54.754 "data_size": 63488 00:14:54.754 }, 00:14:54.754 { 00:14:54.754 "name": "BaseBdev3", 00:14:54.754 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:54.754 "is_configured": true, 00:14:54.754 "data_offset": 2048, 00:14:54.754 "data_size": 63488 00:14:54.754 }, 00:14:54.754 { 00:14:54.754 "name": "BaseBdev4", 00:14:54.754 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:54.754 "is_configured": true, 00:14:54.754 "data_offset": 2048, 00:14:54.754 "data_size": 63488 00:14:54.754 } 00:14:54.754 ] 00:14:54.754 }' 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.754 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.322 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:55.322 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.322 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:55.322 [2024-09-29 16:53:16.807535] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:55.322 [2024-09-29 16:53:16.810867] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000270a0 00:14:55.322 [2024-09-29 16:53:16.813193] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:55.322 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.322 16:53:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:56.259 "name": "raid_bdev1", 00:14:56.259 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:56.259 "strip_size_kb": 64, 00:14:56.259 "state": "online", 00:14:56.259 "raid_level": "raid5f", 00:14:56.259 "superblock": true, 00:14:56.259 "num_base_bdevs": 4, 00:14:56.259 "num_base_bdevs_discovered": 4, 00:14:56.259 "num_base_bdevs_operational": 4, 00:14:56.259 "process": { 00:14:56.259 "type": "rebuild", 00:14:56.259 "target": "spare", 00:14:56.259 "progress": { 00:14:56.259 "blocks": 19200, 00:14:56.259 "percent": 10 00:14:56.259 } 00:14:56.259 }, 00:14:56.259 "base_bdevs_list": [ 00:14:56.259 { 00:14:56.259 "name": "spare", 00:14:56.259 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:14:56.259 "is_configured": true, 00:14:56.259 "data_offset": 2048, 00:14:56.259 "data_size": 63488 00:14:56.259 }, 00:14:56.259 { 00:14:56.259 "name": "BaseBdev2", 00:14:56.259 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:56.259 "is_configured": true, 00:14:56.259 "data_offset": 2048, 00:14:56.259 "data_size": 63488 00:14:56.259 }, 00:14:56.259 { 00:14:56.259 "name": "BaseBdev3", 00:14:56.259 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:56.259 "is_configured": true, 00:14:56.259 "data_offset": 2048, 00:14:56.259 "data_size": 63488 00:14:56.259 }, 00:14:56.259 { 00:14:56.259 "name": "BaseBdev4", 00:14:56.259 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:56.259 "is_configured": true, 00:14:56.259 "data_offset": 2048, 00:14:56.259 "data_size": 63488 00:14:56.259 } 00:14:56.259 ] 00:14:56.259 }' 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:56.259 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:56.519 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:56.519 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:56.519 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.519 16:53:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.519 [2024-09-29 16:53:17.971877] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:56.519 [2024-09-29 16:53:18.018538] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:56.519 [2024-09-29 16:53:18.018603] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:56.519 [2024-09-29 16:53:18.018622] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:56.519 [2024-09-29 16:53:18.018632] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.519 "name": "raid_bdev1", 00:14:56.519 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:56.519 "strip_size_kb": 64, 00:14:56.519 "state": "online", 00:14:56.519 "raid_level": "raid5f", 00:14:56.519 "superblock": true, 00:14:56.519 "num_base_bdevs": 4, 00:14:56.519 "num_base_bdevs_discovered": 3, 00:14:56.519 "num_base_bdevs_operational": 3, 00:14:56.519 "base_bdevs_list": [ 00:14:56.519 { 00:14:56.519 "name": null, 00:14:56.519 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.519 "is_configured": false, 00:14:56.519 "data_offset": 0, 00:14:56.519 "data_size": 63488 00:14:56.519 }, 00:14:56.519 { 00:14:56.519 "name": "BaseBdev2", 00:14:56.519 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:56.519 "is_configured": true, 00:14:56.519 "data_offset": 2048, 00:14:56.519 "data_size": 63488 00:14:56.519 }, 00:14:56.519 { 00:14:56.519 "name": "BaseBdev3", 00:14:56.519 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:56.519 "is_configured": true, 00:14:56.519 "data_offset": 2048, 00:14:56.519 "data_size": 63488 00:14:56.519 }, 00:14:56.519 { 00:14:56.519 "name": "BaseBdev4", 00:14:56.519 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:56.519 "is_configured": true, 00:14:56.519 "data_offset": 2048, 00:14:56.519 "data_size": 63488 00:14:56.519 } 00:14:56.519 ] 00:14:56.519 }' 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.519 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:56.778 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:56.778 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:56.779 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:56.779 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:56.779 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:56.779 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.779 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:56.779 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:57.038 "name": "raid_bdev1", 00:14:57.038 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:57.038 "strip_size_kb": 64, 00:14:57.038 "state": "online", 00:14:57.038 "raid_level": "raid5f", 00:14:57.038 "superblock": true, 00:14:57.038 "num_base_bdevs": 4, 00:14:57.038 "num_base_bdevs_discovered": 3, 00:14:57.038 "num_base_bdevs_operational": 3, 00:14:57.038 "base_bdevs_list": [ 00:14:57.038 { 00:14:57.038 "name": null, 00:14:57.038 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.038 "is_configured": false, 00:14:57.038 "data_offset": 0, 00:14:57.038 "data_size": 63488 00:14:57.038 }, 00:14:57.038 { 00:14:57.038 "name": "BaseBdev2", 00:14:57.038 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:57.038 "is_configured": true, 00:14:57.038 "data_offset": 2048, 00:14:57.038 "data_size": 63488 00:14:57.038 }, 00:14:57.038 { 00:14:57.038 "name": "BaseBdev3", 00:14:57.038 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:57.038 "is_configured": true, 00:14:57.038 "data_offset": 2048, 00:14:57.038 "data_size": 63488 00:14:57.038 }, 00:14:57.038 { 00:14:57.038 "name": "BaseBdev4", 00:14:57.038 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:57.038 "is_configured": true, 00:14:57.038 "data_offset": 2048, 00:14:57.038 "data_size": 63488 00:14:57.038 } 00:14:57.038 ] 00:14:57.038 }' 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.038 [2024-09-29 16:53:18.582913] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:57.038 [2024-09-29 16:53:18.586083] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027170 00:14:57.038 [2024-09-29 16:53:18.588469] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.038 16:53:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:57.977 "name": "raid_bdev1", 00:14:57.977 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:57.977 "strip_size_kb": 64, 00:14:57.977 "state": "online", 00:14:57.977 "raid_level": "raid5f", 00:14:57.977 "superblock": true, 00:14:57.977 "num_base_bdevs": 4, 00:14:57.977 "num_base_bdevs_discovered": 4, 00:14:57.977 "num_base_bdevs_operational": 4, 00:14:57.977 "process": { 00:14:57.977 "type": "rebuild", 00:14:57.977 "target": "spare", 00:14:57.977 "progress": { 00:14:57.977 "blocks": 19200, 00:14:57.977 "percent": 10 00:14:57.977 } 00:14:57.977 }, 00:14:57.977 "base_bdevs_list": [ 00:14:57.977 { 00:14:57.977 "name": "spare", 00:14:57.977 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:14:57.977 "is_configured": true, 00:14:57.977 "data_offset": 2048, 00:14:57.977 "data_size": 63488 00:14:57.977 }, 00:14:57.977 { 00:14:57.977 "name": "BaseBdev2", 00:14:57.977 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:57.977 "is_configured": true, 00:14:57.977 "data_offset": 2048, 00:14:57.977 "data_size": 63488 00:14:57.977 }, 00:14:57.977 { 00:14:57.977 "name": "BaseBdev3", 00:14:57.977 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:57.977 "is_configured": true, 00:14:57.977 "data_offset": 2048, 00:14:57.977 "data_size": 63488 00:14:57.977 }, 00:14:57.977 { 00:14:57.977 "name": "BaseBdev4", 00:14:57.977 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:57.977 "is_configured": true, 00:14:57.977 "data_offset": 2048, 00:14:57.977 "data_size": 63488 00:14:57.977 } 00:14:57.977 ] 00:14:57.977 }' 00:14:57.977 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:58.237 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=524 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:58.237 "name": "raid_bdev1", 00:14:58.237 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:58.237 "strip_size_kb": 64, 00:14:58.237 "state": "online", 00:14:58.237 "raid_level": "raid5f", 00:14:58.237 "superblock": true, 00:14:58.237 "num_base_bdevs": 4, 00:14:58.237 "num_base_bdevs_discovered": 4, 00:14:58.237 "num_base_bdevs_operational": 4, 00:14:58.237 "process": { 00:14:58.237 "type": "rebuild", 00:14:58.237 "target": "spare", 00:14:58.237 "progress": { 00:14:58.237 "blocks": 21120, 00:14:58.237 "percent": 11 00:14:58.237 } 00:14:58.237 }, 00:14:58.237 "base_bdevs_list": [ 00:14:58.237 { 00:14:58.237 "name": "spare", 00:14:58.237 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:14:58.237 "is_configured": true, 00:14:58.237 "data_offset": 2048, 00:14:58.237 "data_size": 63488 00:14:58.237 }, 00:14:58.237 { 00:14:58.237 "name": "BaseBdev2", 00:14:58.237 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:58.237 "is_configured": true, 00:14:58.237 "data_offset": 2048, 00:14:58.237 "data_size": 63488 00:14:58.237 }, 00:14:58.237 { 00:14:58.237 "name": "BaseBdev3", 00:14:58.237 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:58.237 "is_configured": true, 00:14:58.237 "data_offset": 2048, 00:14:58.237 "data_size": 63488 00:14:58.237 }, 00:14:58.237 { 00:14:58.237 "name": "BaseBdev4", 00:14:58.237 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:58.237 "is_configured": true, 00:14:58.237 "data_offset": 2048, 00:14:58.237 "data_size": 63488 00:14:58.237 } 00:14:58.237 ] 00:14:58.237 }' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:58.237 16:53:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:59.619 "name": "raid_bdev1", 00:14:59.619 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:14:59.619 "strip_size_kb": 64, 00:14:59.619 "state": "online", 00:14:59.619 "raid_level": "raid5f", 00:14:59.619 "superblock": true, 00:14:59.619 "num_base_bdevs": 4, 00:14:59.619 "num_base_bdevs_discovered": 4, 00:14:59.619 "num_base_bdevs_operational": 4, 00:14:59.619 "process": { 00:14:59.619 "type": "rebuild", 00:14:59.619 "target": "spare", 00:14:59.619 "progress": { 00:14:59.619 "blocks": 42240, 00:14:59.619 "percent": 22 00:14:59.619 } 00:14:59.619 }, 00:14:59.619 "base_bdevs_list": [ 00:14:59.619 { 00:14:59.619 "name": "spare", 00:14:59.619 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:14:59.619 "is_configured": true, 00:14:59.619 "data_offset": 2048, 00:14:59.619 "data_size": 63488 00:14:59.619 }, 00:14:59.619 { 00:14:59.619 "name": "BaseBdev2", 00:14:59.619 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:14:59.619 "is_configured": true, 00:14:59.619 "data_offset": 2048, 00:14:59.619 "data_size": 63488 00:14:59.619 }, 00:14:59.619 { 00:14:59.619 "name": "BaseBdev3", 00:14:59.619 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:14:59.619 "is_configured": true, 00:14:59.619 "data_offset": 2048, 00:14:59.619 "data_size": 63488 00:14:59.619 }, 00:14:59.619 { 00:14:59.619 "name": "BaseBdev4", 00:14:59.619 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:14:59.619 "is_configured": true, 00:14:59.619 "data_offset": 2048, 00:14:59.619 "data_size": 63488 00:14:59.619 } 00:14:59.619 ] 00:14:59.619 }' 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:59.619 16:53:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.559 16:53:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.559 16:53:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.559 16:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:00.559 "name": "raid_bdev1", 00:15:00.559 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:00.559 "strip_size_kb": 64, 00:15:00.559 "state": "online", 00:15:00.559 "raid_level": "raid5f", 00:15:00.559 "superblock": true, 00:15:00.559 "num_base_bdevs": 4, 00:15:00.559 "num_base_bdevs_discovered": 4, 00:15:00.559 "num_base_bdevs_operational": 4, 00:15:00.559 "process": { 00:15:00.559 "type": "rebuild", 00:15:00.559 "target": "spare", 00:15:00.559 "progress": { 00:15:00.559 "blocks": 63360, 00:15:00.559 "percent": 33 00:15:00.559 } 00:15:00.559 }, 00:15:00.559 "base_bdevs_list": [ 00:15:00.559 { 00:15:00.559 "name": "spare", 00:15:00.559 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:00.559 "is_configured": true, 00:15:00.559 "data_offset": 2048, 00:15:00.559 "data_size": 63488 00:15:00.559 }, 00:15:00.559 { 00:15:00.559 "name": "BaseBdev2", 00:15:00.559 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:00.559 "is_configured": true, 00:15:00.559 "data_offset": 2048, 00:15:00.559 "data_size": 63488 00:15:00.559 }, 00:15:00.559 { 00:15:00.559 "name": "BaseBdev3", 00:15:00.559 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:00.559 "is_configured": true, 00:15:00.559 "data_offset": 2048, 00:15:00.559 "data_size": 63488 00:15:00.559 }, 00:15:00.559 { 00:15:00.559 "name": "BaseBdev4", 00:15:00.559 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:00.559 "is_configured": true, 00:15:00.559 "data_offset": 2048, 00:15:00.559 "data_size": 63488 00:15:00.559 } 00:15:00.559 ] 00:15:00.559 }' 00:15:00.559 16:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:00.559 16:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:00.559 16:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:00.559 16:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:00.559 16:53:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.499 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.758 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.758 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:01.758 "name": "raid_bdev1", 00:15:01.758 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:01.758 "strip_size_kb": 64, 00:15:01.758 "state": "online", 00:15:01.758 "raid_level": "raid5f", 00:15:01.758 "superblock": true, 00:15:01.758 "num_base_bdevs": 4, 00:15:01.758 "num_base_bdevs_discovered": 4, 00:15:01.758 "num_base_bdevs_operational": 4, 00:15:01.758 "process": { 00:15:01.758 "type": "rebuild", 00:15:01.759 "target": "spare", 00:15:01.759 "progress": { 00:15:01.759 "blocks": 86400, 00:15:01.759 "percent": 45 00:15:01.759 } 00:15:01.759 }, 00:15:01.759 "base_bdevs_list": [ 00:15:01.759 { 00:15:01.759 "name": "spare", 00:15:01.759 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:01.759 "is_configured": true, 00:15:01.759 "data_offset": 2048, 00:15:01.759 "data_size": 63488 00:15:01.759 }, 00:15:01.759 { 00:15:01.759 "name": "BaseBdev2", 00:15:01.759 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:01.759 "is_configured": true, 00:15:01.759 "data_offset": 2048, 00:15:01.759 "data_size": 63488 00:15:01.759 }, 00:15:01.759 { 00:15:01.759 "name": "BaseBdev3", 00:15:01.759 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:01.759 "is_configured": true, 00:15:01.759 "data_offset": 2048, 00:15:01.759 "data_size": 63488 00:15:01.759 }, 00:15:01.759 { 00:15:01.759 "name": "BaseBdev4", 00:15:01.759 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:01.759 "is_configured": true, 00:15:01.759 "data_offset": 2048, 00:15:01.759 "data_size": 63488 00:15:01.759 } 00:15:01.759 ] 00:15:01.759 }' 00:15:01.759 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:01.759 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:01.759 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:01.759 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:01.759 16:53:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:02.699 "name": "raid_bdev1", 00:15:02.699 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:02.699 "strip_size_kb": 64, 00:15:02.699 "state": "online", 00:15:02.699 "raid_level": "raid5f", 00:15:02.699 "superblock": true, 00:15:02.699 "num_base_bdevs": 4, 00:15:02.699 "num_base_bdevs_discovered": 4, 00:15:02.699 "num_base_bdevs_operational": 4, 00:15:02.699 "process": { 00:15:02.699 "type": "rebuild", 00:15:02.699 "target": "spare", 00:15:02.699 "progress": { 00:15:02.699 "blocks": 109440, 00:15:02.699 "percent": 57 00:15:02.699 } 00:15:02.699 }, 00:15:02.699 "base_bdevs_list": [ 00:15:02.699 { 00:15:02.699 "name": "spare", 00:15:02.699 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:02.699 "is_configured": true, 00:15:02.699 "data_offset": 2048, 00:15:02.699 "data_size": 63488 00:15:02.699 }, 00:15:02.699 { 00:15:02.699 "name": "BaseBdev2", 00:15:02.699 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:02.699 "is_configured": true, 00:15:02.699 "data_offset": 2048, 00:15:02.699 "data_size": 63488 00:15:02.699 }, 00:15:02.699 { 00:15:02.699 "name": "BaseBdev3", 00:15:02.699 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:02.699 "is_configured": true, 00:15:02.699 "data_offset": 2048, 00:15:02.699 "data_size": 63488 00:15:02.699 }, 00:15:02.699 { 00:15:02.699 "name": "BaseBdev4", 00:15:02.699 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:02.699 "is_configured": true, 00:15:02.699 "data_offset": 2048, 00:15:02.699 "data_size": 63488 00:15:02.699 } 00:15:02.699 ] 00:15:02.699 }' 00:15:02.699 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:02.959 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:02.959 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:02.959 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:02.959 16:53:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:03.898 "name": "raid_bdev1", 00:15:03.898 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:03.898 "strip_size_kb": 64, 00:15:03.898 "state": "online", 00:15:03.898 "raid_level": "raid5f", 00:15:03.898 "superblock": true, 00:15:03.898 "num_base_bdevs": 4, 00:15:03.898 "num_base_bdevs_discovered": 4, 00:15:03.898 "num_base_bdevs_operational": 4, 00:15:03.898 "process": { 00:15:03.898 "type": "rebuild", 00:15:03.898 "target": "spare", 00:15:03.898 "progress": { 00:15:03.898 "blocks": 130560, 00:15:03.898 "percent": 68 00:15:03.898 } 00:15:03.898 }, 00:15:03.898 "base_bdevs_list": [ 00:15:03.898 { 00:15:03.898 "name": "spare", 00:15:03.898 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:03.898 "is_configured": true, 00:15:03.898 "data_offset": 2048, 00:15:03.898 "data_size": 63488 00:15:03.898 }, 00:15:03.898 { 00:15:03.898 "name": "BaseBdev2", 00:15:03.898 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:03.898 "is_configured": true, 00:15:03.898 "data_offset": 2048, 00:15:03.898 "data_size": 63488 00:15:03.898 }, 00:15:03.898 { 00:15:03.898 "name": "BaseBdev3", 00:15:03.898 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:03.898 "is_configured": true, 00:15:03.898 "data_offset": 2048, 00:15:03.898 "data_size": 63488 00:15:03.898 }, 00:15:03.898 { 00:15:03.898 "name": "BaseBdev4", 00:15:03.898 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:03.898 "is_configured": true, 00:15:03.898 "data_offset": 2048, 00:15:03.898 "data_size": 63488 00:15:03.898 } 00:15:03.898 ] 00:15:03.898 }' 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:03.898 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:04.158 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:04.158 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:04.158 16:53:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:05.096 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:05.096 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:05.096 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:05.097 "name": "raid_bdev1", 00:15:05.097 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:05.097 "strip_size_kb": 64, 00:15:05.097 "state": "online", 00:15:05.097 "raid_level": "raid5f", 00:15:05.097 "superblock": true, 00:15:05.097 "num_base_bdevs": 4, 00:15:05.097 "num_base_bdevs_discovered": 4, 00:15:05.097 "num_base_bdevs_operational": 4, 00:15:05.097 "process": { 00:15:05.097 "type": "rebuild", 00:15:05.097 "target": "spare", 00:15:05.097 "progress": { 00:15:05.097 "blocks": 153600, 00:15:05.097 "percent": 80 00:15:05.097 } 00:15:05.097 }, 00:15:05.097 "base_bdevs_list": [ 00:15:05.097 { 00:15:05.097 "name": "spare", 00:15:05.097 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:05.097 "is_configured": true, 00:15:05.097 "data_offset": 2048, 00:15:05.097 "data_size": 63488 00:15:05.097 }, 00:15:05.097 { 00:15:05.097 "name": "BaseBdev2", 00:15:05.097 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:05.097 "is_configured": true, 00:15:05.097 "data_offset": 2048, 00:15:05.097 "data_size": 63488 00:15:05.097 }, 00:15:05.097 { 00:15:05.097 "name": "BaseBdev3", 00:15:05.097 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:05.097 "is_configured": true, 00:15:05.097 "data_offset": 2048, 00:15:05.097 "data_size": 63488 00:15:05.097 }, 00:15:05.097 { 00:15:05.097 "name": "BaseBdev4", 00:15:05.097 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:05.097 "is_configured": true, 00:15:05.097 "data_offset": 2048, 00:15:05.097 "data_size": 63488 00:15:05.097 } 00:15:05.097 ] 00:15:05.097 }' 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:05.097 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:05.357 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:05.357 16:53:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:06.296 "name": "raid_bdev1", 00:15:06.296 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:06.296 "strip_size_kb": 64, 00:15:06.296 "state": "online", 00:15:06.296 "raid_level": "raid5f", 00:15:06.296 "superblock": true, 00:15:06.296 "num_base_bdevs": 4, 00:15:06.296 "num_base_bdevs_discovered": 4, 00:15:06.296 "num_base_bdevs_operational": 4, 00:15:06.296 "process": { 00:15:06.296 "type": "rebuild", 00:15:06.296 "target": "spare", 00:15:06.296 "progress": { 00:15:06.296 "blocks": 174720, 00:15:06.296 "percent": 91 00:15:06.296 } 00:15:06.296 }, 00:15:06.296 "base_bdevs_list": [ 00:15:06.296 { 00:15:06.296 "name": "spare", 00:15:06.296 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:06.296 "is_configured": true, 00:15:06.296 "data_offset": 2048, 00:15:06.296 "data_size": 63488 00:15:06.296 }, 00:15:06.296 { 00:15:06.296 "name": "BaseBdev2", 00:15:06.296 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:06.296 "is_configured": true, 00:15:06.296 "data_offset": 2048, 00:15:06.296 "data_size": 63488 00:15:06.296 }, 00:15:06.296 { 00:15:06.296 "name": "BaseBdev3", 00:15:06.296 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:06.296 "is_configured": true, 00:15:06.296 "data_offset": 2048, 00:15:06.296 "data_size": 63488 00:15:06.296 }, 00:15:06.296 { 00:15:06.296 "name": "BaseBdev4", 00:15:06.296 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:06.296 "is_configured": true, 00:15:06.296 "data_offset": 2048, 00:15:06.296 "data_size": 63488 00:15:06.296 } 00:15:06.296 ] 00:15:06.296 }' 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:06.296 16:53:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:07.234 [2024-09-29 16:53:28.627811] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:07.234 [2024-09-29 16:53:28.627881] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:07.234 [2024-09-29 16:53:28.628009] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.492 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.492 "name": "raid_bdev1", 00:15:07.492 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:07.492 "strip_size_kb": 64, 00:15:07.492 "state": "online", 00:15:07.492 "raid_level": "raid5f", 00:15:07.492 "superblock": true, 00:15:07.492 "num_base_bdevs": 4, 00:15:07.492 "num_base_bdevs_discovered": 4, 00:15:07.492 "num_base_bdevs_operational": 4, 00:15:07.492 "base_bdevs_list": [ 00:15:07.492 { 00:15:07.492 "name": "spare", 00:15:07.492 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:07.492 "is_configured": true, 00:15:07.492 "data_offset": 2048, 00:15:07.492 "data_size": 63488 00:15:07.492 }, 00:15:07.492 { 00:15:07.492 "name": "BaseBdev2", 00:15:07.492 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:07.492 "is_configured": true, 00:15:07.492 "data_offset": 2048, 00:15:07.492 "data_size": 63488 00:15:07.492 }, 00:15:07.492 { 00:15:07.492 "name": "BaseBdev3", 00:15:07.492 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:07.492 "is_configured": true, 00:15:07.493 "data_offset": 2048, 00:15:07.493 "data_size": 63488 00:15:07.493 }, 00:15:07.493 { 00:15:07.493 "name": "BaseBdev4", 00:15:07.493 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:07.493 "is_configured": true, 00:15:07.493 "data_offset": 2048, 00:15:07.493 "data_size": 63488 00:15:07.493 } 00:15:07.493 ] 00:15:07.493 }' 00:15:07.493 16:53:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:07.493 "name": "raid_bdev1", 00:15:07.493 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:07.493 "strip_size_kb": 64, 00:15:07.493 "state": "online", 00:15:07.493 "raid_level": "raid5f", 00:15:07.493 "superblock": true, 00:15:07.493 "num_base_bdevs": 4, 00:15:07.493 "num_base_bdevs_discovered": 4, 00:15:07.493 "num_base_bdevs_operational": 4, 00:15:07.493 "base_bdevs_list": [ 00:15:07.493 { 00:15:07.493 "name": "spare", 00:15:07.493 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:07.493 "is_configured": true, 00:15:07.493 "data_offset": 2048, 00:15:07.493 "data_size": 63488 00:15:07.493 }, 00:15:07.493 { 00:15:07.493 "name": "BaseBdev2", 00:15:07.493 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:07.493 "is_configured": true, 00:15:07.493 "data_offset": 2048, 00:15:07.493 "data_size": 63488 00:15:07.493 }, 00:15:07.493 { 00:15:07.493 "name": "BaseBdev3", 00:15:07.493 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:07.493 "is_configured": true, 00:15:07.493 "data_offset": 2048, 00:15:07.493 "data_size": 63488 00:15:07.493 }, 00:15:07.493 { 00:15:07.493 "name": "BaseBdev4", 00:15:07.493 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:07.493 "is_configured": true, 00:15:07.493 "data_offset": 2048, 00:15:07.493 "data_size": 63488 00:15:07.493 } 00:15:07.493 ] 00:15:07.493 }' 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:07.493 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:07.752 "name": "raid_bdev1", 00:15:07.752 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:07.752 "strip_size_kb": 64, 00:15:07.752 "state": "online", 00:15:07.752 "raid_level": "raid5f", 00:15:07.752 "superblock": true, 00:15:07.752 "num_base_bdevs": 4, 00:15:07.752 "num_base_bdevs_discovered": 4, 00:15:07.752 "num_base_bdevs_operational": 4, 00:15:07.752 "base_bdevs_list": [ 00:15:07.752 { 00:15:07.752 "name": "spare", 00:15:07.752 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:07.752 "is_configured": true, 00:15:07.752 "data_offset": 2048, 00:15:07.752 "data_size": 63488 00:15:07.752 }, 00:15:07.752 { 00:15:07.752 "name": "BaseBdev2", 00:15:07.752 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:07.752 "is_configured": true, 00:15:07.752 "data_offset": 2048, 00:15:07.752 "data_size": 63488 00:15:07.752 }, 00:15:07.752 { 00:15:07.752 "name": "BaseBdev3", 00:15:07.752 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:07.752 "is_configured": true, 00:15:07.752 "data_offset": 2048, 00:15:07.752 "data_size": 63488 00:15:07.752 }, 00:15:07.752 { 00:15:07.752 "name": "BaseBdev4", 00:15:07.752 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:07.752 "is_configured": true, 00:15:07.752 "data_offset": 2048, 00:15:07.752 "data_size": 63488 00:15:07.752 } 00:15:07.752 ] 00:15:07.752 }' 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:07.752 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.012 [2024-09-29 16:53:29.615408] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:08.012 [2024-09-29 16:53:29.615443] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:08.012 [2024-09-29 16:53:29.615512] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:08.012 [2024-09-29 16:53:29.615676] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:08.012 [2024-09-29 16:53:29.615710] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:08.012 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:08.271 /dev/nbd0 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:08.271 1+0 records in 00:15:08.271 1+0 records out 00:15:08.271 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000483935 s, 8.5 MB/s 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:15:08.271 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:08.272 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:08.272 16:53:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:08.531 /dev/nbd1 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:08.531 1+0 records in 00:15:08.531 1+0 records out 00:15:08.531 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000402821 s, 10.2 MB/s 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:08.531 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:08.791 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.051 [2024-09-29 16:53:30.702611] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:09.051 [2024-09-29 16:53:30.702684] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:09.051 [2024-09-29 16:53:30.702707] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:09.051 [2024-09-29 16:53:30.702730] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:09.051 [2024-09-29 16:53:30.704937] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:09.051 [2024-09-29 16:53:30.704982] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:09.051 [2024-09-29 16:53:30.705063] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:09.051 [2024-09-29 16:53:30.705113] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:09.051 [2024-09-29 16:53:30.705234] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:09.051 [2024-09-29 16:53:30.705360] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:09.051 [2024-09-29 16:53:30.705448] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:09.051 spare 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.051 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.311 [2024-09-29 16:53:30.805360] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:09.311 [2024-09-29 16:53:30.805389] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:09.311 [2024-09-29 16:53:30.805648] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000045820 00:15:09.311 [2024-09-29 16:53:30.806143] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:09.311 [2024-09-29 16:53:30.806175] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:09.311 [2024-09-29 16:53:30.806312] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:09.311 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.311 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.312 "name": "raid_bdev1", 00:15:09.312 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:09.312 "strip_size_kb": 64, 00:15:09.312 "state": "online", 00:15:09.312 "raid_level": "raid5f", 00:15:09.312 "superblock": true, 00:15:09.312 "num_base_bdevs": 4, 00:15:09.312 "num_base_bdevs_discovered": 4, 00:15:09.312 "num_base_bdevs_operational": 4, 00:15:09.312 "base_bdevs_list": [ 00:15:09.312 { 00:15:09.312 "name": "spare", 00:15:09.312 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:09.312 "is_configured": true, 00:15:09.312 "data_offset": 2048, 00:15:09.312 "data_size": 63488 00:15:09.312 }, 00:15:09.312 { 00:15:09.312 "name": "BaseBdev2", 00:15:09.312 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:09.312 "is_configured": true, 00:15:09.312 "data_offset": 2048, 00:15:09.312 "data_size": 63488 00:15:09.312 }, 00:15:09.312 { 00:15:09.312 "name": "BaseBdev3", 00:15:09.312 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:09.312 "is_configured": true, 00:15:09.312 "data_offset": 2048, 00:15:09.312 "data_size": 63488 00:15:09.312 }, 00:15:09.312 { 00:15:09.312 "name": "BaseBdev4", 00:15:09.312 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:09.312 "is_configured": true, 00:15:09.312 "data_offset": 2048, 00:15:09.312 "data_size": 63488 00:15:09.312 } 00:15:09.312 ] 00:15:09.312 }' 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.312 16:53:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.571 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.831 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.831 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.831 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:09.831 "name": "raid_bdev1", 00:15:09.831 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:09.831 "strip_size_kb": 64, 00:15:09.831 "state": "online", 00:15:09.831 "raid_level": "raid5f", 00:15:09.831 "superblock": true, 00:15:09.831 "num_base_bdevs": 4, 00:15:09.831 "num_base_bdevs_discovered": 4, 00:15:09.831 "num_base_bdevs_operational": 4, 00:15:09.832 "base_bdevs_list": [ 00:15:09.832 { 00:15:09.832 "name": "spare", 00:15:09.832 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:09.832 "is_configured": true, 00:15:09.832 "data_offset": 2048, 00:15:09.832 "data_size": 63488 00:15:09.832 }, 00:15:09.832 { 00:15:09.832 "name": "BaseBdev2", 00:15:09.832 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:09.832 "is_configured": true, 00:15:09.832 "data_offset": 2048, 00:15:09.832 "data_size": 63488 00:15:09.832 }, 00:15:09.832 { 00:15:09.832 "name": "BaseBdev3", 00:15:09.832 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:09.832 "is_configured": true, 00:15:09.832 "data_offset": 2048, 00:15:09.832 "data_size": 63488 00:15:09.832 }, 00:15:09.832 { 00:15:09.832 "name": "BaseBdev4", 00:15:09.832 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:09.832 "is_configured": true, 00:15:09.832 "data_offset": 2048, 00:15:09.832 "data_size": 63488 00:15:09.832 } 00:15:09.832 ] 00:15:09.832 }' 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.832 [2024-09-29 16:53:31.419015] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:09.832 "name": "raid_bdev1", 00:15:09.832 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:09.832 "strip_size_kb": 64, 00:15:09.832 "state": "online", 00:15:09.832 "raid_level": "raid5f", 00:15:09.832 "superblock": true, 00:15:09.832 "num_base_bdevs": 4, 00:15:09.832 "num_base_bdevs_discovered": 3, 00:15:09.832 "num_base_bdevs_operational": 3, 00:15:09.832 "base_bdevs_list": [ 00:15:09.832 { 00:15:09.832 "name": null, 00:15:09.832 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:09.832 "is_configured": false, 00:15:09.832 "data_offset": 0, 00:15:09.832 "data_size": 63488 00:15:09.832 }, 00:15:09.832 { 00:15:09.832 "name": "BaseBdev2", 00:15:09.832 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:09.832 "is_configured": true, 00:15:09.832 "data_offset": 2048, 00:15:09.832 "data_size": 63488 00:15:09.832 }, 00:15:09.832 { 00:15:09.832 "name": "BaseBdev3", 00:15:09.832 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:09.832 "is_configured": true, 00:15:09.832 "data_offset": 2048, 00:15:09.832 "data_size": 63488 00:15:09.832 }, 00:15:09.832 { 00:15:09.832 "name": "BaseBdev4", 00:15:09.832 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:09.832 "is_configured": true, 00:15:09.832 "data_offset": 2048, 00:15:09.832 "data_size": 63488 00:15:09.832 } 00:15:09.832 ] 00:15:09.832 }' 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:09.832 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.402 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:10.402 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.402 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:10.402 [2024-09-29 16:53:31.850318] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:10.402 [2024-09-29 16:53:31.850441] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:10.402 [2024-09-29 16:53:31.850478] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:10.402 [2024-09-29 16:53:31.850520] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:10.402 [2024-09-29 16:53:31.853742] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000458f0 00:15:10.402 [2024-09-29 16:53:31.855979] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:10.402 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.402 16:53:31 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:11.341 "name": "raid_bdev1", 00:15:11.341 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:11.341 "strip_size_kb": 64, 00:15:11.341 "state": "online", 00:15:11.341 "raid_level": "raid5f", 00:15:11.341 "superblock": true, 00:15:11.341 "num_base_bdevs": 4, 00:15:11.341 "num_base_bdevs_discovered": 4, 00:15:11.341 "num_base_bdevs_operational": 4, 00:15:11.341 "process": { 00:15:11.341 "type": "rebuild", 00:15:11.341 "target": "spare", 00:15:11.341 "progress": { 00:15:11.341 "blocks": 19200, 00:15:11.341 "percent": 10 00:15:11.341 } 00:15:11.341 }, 00:15:11.341 "base_bdevs_list": [ 00:15:11.341 { 00:15:11.341 "name": "spare", 00:15:11.341 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:11.341 "is_configured": true, 00:15:11.341 "data_offset": 2048, 00:15:11.341 "data_size": 63488 00:15:11.341 }, 00:15:11.341 { 00:15:11.341 "name": "BaseBdev2", 00:15:11.341 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:11.341 "is_configured": true, 00:15:11.341 "data_offset": 2048, 00:15:11.341 "data_size": 63488 00:15:11.341 }, 00:15:11.341 { 00:15:11.341 "name": "BaseBdev3", 00:15:11.341 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:11.341 "is_configured": true, 00:15:11.341 "data_offset": 2048, 00:15:11.341 "data_size": 63488 00:15:11.341 }, 00:15:11.341 { 00:15:11.341 "name": "BaseBdev4", 00:15:11.341 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:11.341 "is_configured": true, 00:15:11.341 "data_offset": 2048, 00:15:11.341 "data_size": 63488 00:15:11.341 } 00:15:11.341 ] 00:15:11.341 }' 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.341 16:53:32 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.341 [2024-09-29 16:53:32.999397] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:11.601 [2024-09-29 16:53:33.061101] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:11.601 [2024-09-29 16:53:33.061155] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:11.601 [2024-09-29 16:53:33.061173] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:11.601 [2024-09-29 16:53:33.061181] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:11.601 "name": "raid_bdev1", 00:15:11.601 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:11.601 "strip_size_kb": 64, 00:15:11.601 "state": "online", 00:15:11.601 "raid_level": "raid5f", 00:15:11.601 "superblock": true, 00:15:11.601 "num_base_bdevs": 4, 00:15:11.601 "num_base_bdevs_discovered": 3, 00:15:11.601 "num_base_bdevs_operational": 3, 00:15:11.601 "base_bdevs_list": [ 00:15:11.601 { 00:15:11.601 "name": null, 00:15:11.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:11.601 "is_configured": false, 00:15:11.601 "data_offset": 0, 00:15:11.601 "data_size": 63488 00:15:11.601 }, 00:15:11.601 { 00:15:11.601 "name": "BaseBdev2", 00:15:11.601 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:11.601 "is_configured": true, 00:15:11.601 "data_offset": 2048, 00:15:11.601 "data_size": 63488 00:15:11.601 }, 00:15:11.601 { 00:15:11.601 "name": "BaseBdev3", 00:15:11.601 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:11.601 "is_configured": true, 00:15:11.601 "data_offset": 2048, 00:15:11.601 "data_size": 63488 00:15:11.601 }, 00:15:11.601 { 00:15:11.601 "name": "BaseBdev4", 00:15:11.601 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:11.601 "is_configured": true, 00:15:11.601 "data_offset": 2048, 00:15:11.601 "data_size": 63488 00:15:11.601 } 00:15:11.601 ] 00:15:11.601 }' 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:11.601 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:11.861 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:11.861 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.861 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:12.121 [2024-09-29 16:53:33.537043] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:12.121 [2024-09-29 16:53:33.537161] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:12.121 [2024-09-29 16:53:33.537200] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:15:12.121 [2024-09-29 16:53:33.537229] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:12.121 [2024-09-29 16:53:33.537688] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:12.121 [2024-09-29 16:53:33.537763] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:12.121 [2024-09-29 16:53:33.537888] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:12.121 [2024-09-29 16:53:33.537929] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:12.121 [2024-09-29 16:53:33.537992] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:12.121 [2024-09-29 16:53:33.538042] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:12.121 [2024-09-29 16:53:33.540679] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000459c0 00:15:12.121 [2024-09-29 16:53:33.542867] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:12.121 spare 00:15:12.121 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.121 16:53:33 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:13.059 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:13.059 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:13.059 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:13.059 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:13.059 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:13.059 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:13.060 "name": "raid_bdev1", 00:15:13.060 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:13.060 "strip_size_kb": 64, 00:15:13.060 "state": "online", 00:15:13.060 "raid_level": "raid5f", 00:15:13.060 "superblock": true, 00:15:13.060 "num_base_bdevs": 4, 00:15:13.060 "num_base_bdevs_discovered": 4, 00:15:13.060 "num_base_bdevs_operational": 4, 00:15:13.060 "process": { 00:15:13.060 "type": "rebuild", 00:15:13.060 "target": "spare", 00:15:13.060 "progress": { 00:15:13.060 "blocks": 19200, 00:15:13.060 "percent": 10 00:15:13.060 } 00:15:13.060 }, 00:15:13.060 "base_bdevs_list": [ 00:15:13.060 { 00:15:13.060 "name": "spare", 00:15:13.060 "uuid": "b7219a2a-a2cd-529f-907e-0e66cc5a703e", 00:15:13.060 "is_configured": true, 00:15:13.060 "data_offset": 2048, 00:15:13.060 "data_size": 63488 00:15:13.060 }, 00:15:13.060 { 00:15:13.060 "name": "BaseBdev2", 00:15:13.060 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:13.060 "is_configured": true, 00:15:13.060 "data_offset": 2048, 00:15:13.060 "data_size": 63488 00:15:13.060 }, 00:15:13.060 { 00:15:13.060 "name": "BaseBdev3", 00:15:13.060 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:13.060 "is_configured": true, 00:15:13.060 "data_offset": 2048, 00:15:13.060 "data_size": 63488 00:15:13.060 }, 00:15:13.060 { 00:15:13.060 "name": "BaseBdev4", 00:15:13.060 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:13.060 "is_configured": true, 00:15:13.060 "data_offset": 2048, 00:15:13.060 "data_size": 63488 00:15:13.060 } 00:15:13.060 ] 00:15:13.060 }' 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.060 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.060 [2024-09-29 16:53:34.681583] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:13.320 [2024-09-29 16:53:34.748054] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:13.320 [2024-09-29 16:53:34.748180] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:13.320 [2024-09-29 16:53:34.748218] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:13.320 [2024-09-29 16:53:34.748244] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:13.320 "name": "raid_bdev1", 00:15:13.320 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:13.320 "strip_size_kb": 64, 00:15:13.320 "state": "online", 00:15:13.320 "raid_level": "raid5f", 00:15:13.320 "superblock": true, 00:15:13.320 "num_base_bdevs": 4, 00:15:13.320 "num_base_bdevs_discovered": 3, 00:15:13.320 "num_base_bdevs_operational": 3, 00:15:13.320 "base_bdevs_list": [ 00:15:13.320 { 00:15:13.320 "name": null, 00:15:13.320 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:13.320 "is_configured": false, 00:15:13.320 "data_offset": 0, 00:15:13.320 "data_size": 63488 00:15:13.320 }, 00:15:13.320 { 00:15:13.320 "name": "BaseBdev2", 00:15:13.320 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:13.320 "is_configured": true, 00:15:13.320 "data_offset": 2048, 00:15:13.320 "data_size": 63488 00:15:13.320 }, 00:15:13.320 { 00:15:13.320 "name": "BaseBdev3", 00:15:13.320 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:13.320 "is_configured": true, 00:15:13.320 "data_offset": 2048, 00:15:13.320 "data_size": 63488 00:15:13.320 }, 00:15:13.320 { 00:15:13.320 "name": "BaseBdev4", 00:15:13.320 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:13.320 "is_configured": true, 00:15:13.320 "data_offset": 2048, 00:15:13.320 "data_size": 63488 00:15:13.320 } 00:15:13.320 ] 00:15:13.320 }' 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:13.320 16:53:34 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.579 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:13.579 "name": "raid_bdev1", 00:15:13.579 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:13.579 "strip_size_kb": 64, 00:15:13.579 "state": "online", 00:15:13.579 "raid_level": "raid5f", 00:15:13.579 "superblock": true, 00:15:13.579 "num_base_bdevs": 4, 00:15:13.579 "num_base_bdevs_discovered": 3, 00:15:13.580 "num_base_bdevs_operational": 3, 00:15:13.580 "base_bdevs_list": [ 00:15:13.580 { 00:15:13.580 "name": null, 00:15:13.580 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:13.580 "is_configured": false, 00:15:13.580 "data_offset": 0, 00:15:13.580 "data_size": 63488 00:15:13.580 }, 00:15:13.580 { 00:15:13.580 "name": "BaseBdev2", 00:15:13.580 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:13.580 "is_configured": true, 00:15:13.580 "data_offset": 2048, 00:15:13.580 "data_size": 63488 00:15:13.580 }, 00:15:13.580 { 00:15:13.580 "name": "BaseBdev3", 00:15:13.580 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:13.580 "is_configured": true, 00:15:13.580 "data_offset": 2048, 00:15:13.580 "data_size": 63488 00:15:13.580 }, 00:15:13.580 { 00:15:13.580 "name": "BaseBdev4", 00:15:13.580 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:13.580 "is_configured": true, 00:15:13.580 "data_offset": 2048, 00:15:13.580 "data_size": 63488 00:15:13.580 } 00:15:13.580 ] 00:15:13.580 }' 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:13.843 [2024-09-29 16:53:35.343956] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:13.843 [2024-09-29 16:53:35.344060] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.843 [2024-09-29 16:53:35.344113] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:15:13.843 [2024-09-29 16:53:35.344142] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.843 [2024-09-29 16:53:35.344575] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.843 [2024-09-29 16:53:35.344637] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:13.843 [2024-09-29 16:53:35.344751] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:13.843 [2024-09-29 16:53:35.344799] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:13.843 [2024-09-29 16:53:35.344843] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:13.843 [2024-09-29 16:53:35.344904] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:13.843 BaseBdev1 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.843 16:53:35 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.780 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.780 "name": "raid_bdev1", 00:15:14.780 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:14.780 "strip_size_kb": 64, 00:15:14.780 "state": "online", 00:15:14.780 "raid_level": "raid5f", 00:15:14.780 "superblock": true, 00:15:14.780 "num_base_bdevs": 4, 00:15:14.780 "num_base_bdevs_discovered": 3, 00:15:14.780 "num_base_bdevs_operational": 3, 00:15:14.780 "base_bdevs_list": [ 00:15:14.780 { 00:15:14.780 "name": null, 00:15:14.780 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:14.780 "is_configured": false, 00:15:14.780 "data_offset": 0, 00:15:14.780 "data_size": 63488 00:15:14.780 }, 00:15:14.780 { 00:15:14.780 "name": "BaseBdev2", 00:15:14.780 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:14.780 "is_configured": true, 00:15:14.780 "data_offset": 2048, 00:15:14.780 "data_size": 63488 00:15:14.780 }, 00:15:14.780 { 00:15:14.780 "name": "BaseBdev3", 00:15:14.780 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:14.780 "is_configured": true, 00:15:14.780 "data_offset": 2048, 00:15:14.780 "data_size": 63488 00:15:14.781 }, 00:15:14.781 { 00:15:14.781 "name": "BaseBdev4", 00:15:14.781 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:14.781 "is_configured": true, 00:15:14.781 "data_offset": 2048, 00:15:14.781 "data_size": 63488 00:15:14.781 } 00:15:14.781 ] 00:15:14.781 }' 00:15:14.781 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.781 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:15.349 "name": "raid_bdev1", 00:15:15.349 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:15.349 "strip_size_kb": 64, 00:15:15.349 "state": "online", 00:15:15.349 "raid_level": "raid5f", 00:15:15.349 "superblock": true, 00:15:15.349 "num_base_bdevs": 4, 00:15:15.349 "num_base_bdevs_discovered": 3, 00:15:15.349 "num_base_bdevs_operational": 3, 00:15:15.349 "base_bdevs_list": [ 00:15:15.349 { 00:15:15.349 "name": null, 00:15:15.349 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.349 "is_configured": false, 00:15:15.349 "data_offset": 0, 00:15:15.349 "data_size": 63488 00:15:15.349 }, 00:15:15.349 { 00:15:15.349 "name": "BaseBdev2", 00:15:15.349 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:15.349 "is_configured": true, 00:15:15.349 "data_offset": 2048, 00:15:15.349 "data_size": 63488 00:15:15.349 }, 00:15:15.349 { 00:15:15.349 "name": "BaseBdev3", 00:15:15.349 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:15.349 "is_configured": true, 00:15:15.349 "data_offset": 2048, 00:15:15.349 "data_size": 63488 00:15:15.349 }, 00:15:15.349 { 00:15:15.349 "name": "BaseBdev4", 00:15:15.349 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:15.349 "is_configured": true, 00:15:15.349 "data_offset": 2048, 00:15:15.349 "data_size": 63488 00:15:15.349 } 00:15:15.349 ] 00:15:15.349 }' 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:15.349 [2024-09-29 16:53:36.917309] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:15.349 [2024-09-29 16:53:36.917427] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:15.349 [2024-09-29 16:53:36.917441] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:15.349 request: 00:15:15.349 { 00:15:15.349 "base_bdev": "BaseBdev1", 00:15:15.349 "raid_bdev": "raid_bdev1", 00:15:15.349 "method": "bdev_raid_add_base_bdev", 00:15:15.349 "req_id": 1 00:15:15.349 } 00:15:15.349 Got JSON-RPC error response 00:15:15.349 response: 00:15:15.349 { 00:15:15.349 "code": -22, 00:15:15.349 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:15.349 } 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:15.349 16:53:36 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.287 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.546 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.546 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:16.546 "name": "raid_bdev1", 00:15:16.546 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:16.546 "strip_size_kb": 64, 00:15:16.546 "state": "online", 00:15:16.546 "raid_level": "raid5f", 00:15:16.546 "superblock": true, 00:15:16.546 "num_base_bdevs": 4, 00:15:16.546 "num_base_bdevs_discovered": 3, 00:15:16.546 "num_base_bdevs_operational": 3, 00:15:16.546 "base_bdevs_list": [ 00:15:16.546 { 00:15:16.546 "name": null, 00:15:16.546 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:16.546 "is_configured": false, 00:15:16.546 "data_offset": 0, 00:15:16.546 "data_size": 63488 00:15:16.546 }, 00:15:16.546 { 00:15:16.546 "name": "BaseBdev2", 00:15:16.546 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:16.546 "is_configured": true, 00:15:16.546 "data_offset": 2048, 00:15:16.546 "data_size": 63488 00:15:16.546 }, 00:15:16.546 { 00:15:16.546 "name": "BaseBdev3", 00:15:16.546 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:16.546 "is_configured": true, 00:15:16.546 "data_offset": 2048, 00:15:16.546 "data_size": 63488 00:15:16.546 }, 00:15:16.546 { 00:15:16.546 "name": "BaseBdev4", 00:15:16.546 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:16.546 "is_configured": true, 00:15:16.546 "data_offset": 2048, 00:15:16.546 "data_size": 63488 00:15:16.546 } 00:15:16.546 ] 00:15:16.546 }' 00:15:16.546 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:16.546 16:53:37 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.804 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:16.804 "name": "raid_bdev1", 00:15:16.804 "uuid": "1b5ded75-5f91-4cfa-94b4-69867071d1e8", 00:15:16.804 "strip_size_kb": 64, 00:15:16.804 "state": "online", 00:15:16.804 "raid_level": "raid5f", 00:15:16.804 "superblock": true, 00:15:16.804 "num_base_bdevs": 4, 00:15:16.804 "num_base_bdevs_discovered": 3, 00:15:16.804 "num_base_bdevs_operational": 3, 00:15:16.804 "base_bdevs_list": [ 00:15:16.804 { 00:15:16.804 "name": null, 00:15:16.804 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:16.804 "is_configured": false, 00:15:16.804 "data_offset": 0, 00:15:16.804 "data_size": 63488 00:15:16.805 }, 00:15:16.805 { 00:15:16.805 "name": "BaseBdev2", 00:15:16.805 "uuid": "ce2d2e24-f75d-567a-a614-e434f0775cf3", 00:15:16.805 "is_configured": true, 00:15:16.805 "data_offset": 2048, 00:15:16.805 "data_size": 63488 00:15:16.805 }, 00:15:16.805 { 00:15:16.805 "name": "BaseBdev3", 00:15:16.805 "uuid": "6d8e67c7-5de6-5d78-be94-73688de57e54", 00:15:16.805 "is_configured": true, 00:15:16.805 "data_offset": 2048, 00:15:16.805 "data_size": 63488 00:15:16.805 }, 00:15:16.805 { 00:15:16.805 "name": "BaseBdev4", 00:15:16.805 "uuid": "b02555fc-6dca-5a87-8cb1-cb91c272301e", 00:15:16.805 "is_configured": true, 00:15:16.805 "data_offset": 2048, 00:15:16.805 "data_size": 63488 00:15:16.805 } 00:15:16.805 ] 00:15:16.805 }' 00:15:16.805 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 95152 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 95152 ']' 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 95152 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 95152 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:17.063 killing process with pid 95152 00:15:17.063 Received shutdown signal, test time was about 60.000000 seconds 00:15:17.063 00:15:17.063 Latency(us) 00:15:17.063 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:17.063 =================================================================================================================== 00:15:17.063 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 95152' 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 95152 00:15:17.063 [2024-09-29 16:53:38.597668] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:17.063 [2024-09-29 16:53:38.597775] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:17.063 [2024-09-29 16:53:38.597840] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:17.063 [2024-09-29 16:53:38.597849] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:17.063 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 95152 00:15:17.063 [2024-09-29 16:53:38.647716] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:17.323 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:17.323 00:15:17.323 real 0m25.115s 00:15:17.323 user 0m31.840s 00:15:17.323 sys 0m3.035s 00:15:17.323 ************************************ 00:15:17.323 END TEST raid5f_rebuild_test_sb 00:15:17.323 ************************************ 00:15:17.323 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:17.323 16:53:38 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:17.323 16:53:38 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:15:17.323 16:53:38 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:15:17.323 16:53:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:15:17.323 16:53:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:17.323 16:53:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:17.323 ************************************ 00:15:17.323 START TEST raid_state_function_test_sb_4k 00:15:17.323 ************************************ 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:17.323 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=95950 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 95950' 00:15:17.324 Process raid pid: 95950 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 95950 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 95950 ']' 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:17.324 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:17.324 16:53:38 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:17.583 [2024-09-29 16:53:39.062560] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:15:17.583 [2024-09-29 16:53:39.062792] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:17.583 [2024-09-29 16:53:39.210513] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:17.842 [2024-09-29 16:53:39.256381] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:17.842 [2024-09-29 16:53:39.299276] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.842 [2024-09-29 16:53:39.299389] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.410 [2024-09-29 16:53:39.912878] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:18.410 [2024-09-29 16:53:39.912991] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:18.410 [2024-09-29 16:53:39.913022] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:18.410 [2024-09-29 16:53:39.913044] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.410 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:18.410 "name": "Existed_Raid", 00:15:18.410 "uuid": "34e5231f-45a3-40a3-affa-50089d6892b2", 00:15:18.410 "strip_size_kb": 0, 00:15:18.410 "state": "configuring", 00:15:18.410 "raid_level": "raid1", 00:15:18.411 "superblock": true, 00:15:18.411 "num_base_bdevs": 2, 00:15:18.411 "num_base_bdevs_discovered": 0, 00:15:18.411 "num_base_bdevs_operational": 2, 00:15:18.411 "base_bdevs_list": [ 00:15:18.411 { 00:15:18.411 "name": "BaseBdev1", 00:15:18.411 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:18.411 "is_configured": false, 00:15:18.411 "data_offset": 0, 00:15:18.411 "data_size": 0 00:15:18.411 }, 00:15:18.411 { 00:15:18.411 "name": "BaseBdev2", 00:15:18.411 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:18.411 "is_configured": false, 00:15:18.411 "data_offset": 0, 00:15:18.411 "data_size": 0 00:15:18.411 } 00:15:18.411 ] 00:15:18.411 }' 00:15:18.411 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:18.411 16:53:39 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.979 [2024-09-29 16:53:40.391916] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:18.979 [2024-09-29 16:53:40.391956] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.979 [2024-09-29 16:53:40.399917] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:18.979 [2024-09-29 16:53:40.400014] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:18.979 [2024-09-29 16:53:40.400054] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:18.979 [2024-09-29 16:53:40.400078] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.979 [2024-09-29 16:53:40.417029] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:18.979 BaseBdev1 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.979 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.979 [ 00:15:18.979 { 00:15:18.979 "name": "BaseBdev1", 00:15:18.979 "aliases": [ 00:15:18.979 "5da35ddd-8ba0-4d99-811d-56f9700ee42d" 00:15:18.979 ], 00:15:18.980 "product_name": "Malloc disk", 00:15:18.980 "block_size": 4096, 00:15:18.980 "num_blocks": 8192, 00:15:18.980 "uuid": "5da35ddd-8ba0-4d99-811d-56f9700ee42d", 00:15:18.980 "assigned_rate_limits": { 00:15:18.980 "rw_ios_per_sec": 0, 00:15:18.980 "rw_mbytes_per_sec": 0, 00:15:18.980 "r_mbytes_per_sec": 0, 00:15:18.980 "w_mbytes_per_sec": 0 00:15:18.980 }, 00:15:18.980 "claimed": true, 00:15:18.980 "claim_type": "exclusive_write", 00:15:18.980 "zoned": false, 00:15:18.980 "supported_io_types": { 00:15:18.980 "read": true, 00:15:18.980 "write": true, 00:15:18.980 "unmap": true, 00:15:18.980 "flush": true, 00:15:18.980 "reset": true, 00:15:18.980 "nvme_admin": false, 00:15:18.980 "nvme_io": false, 00:15:18.980 "nvme_io_md": false, 00:15:18.980 "write_zeroes": true, 00:15:18.980 "zcopy": true, 00:15:18.980 "get_zone_info": false, 00:15:18.980 "zone_management": false, 00:15:18.980 "zone_append": false, 00:15:18.980 "compare": false, 00:15:18.980 "compare_and_write": false, 00:15:18.980 "abort": true, 00:15:18.980 "seek_hole": false, 00:15:18.980 "seek_data": false, 00:15:18.980 "copy": true, 00:15:18.980 "nvme_iov_md": false 00:15:18.980 }, 00:15:18.980 "memory_domains": [ 00:15:18.980 { 00:15:18.980 "dma_device_id": "system", 00:15:18.980 "dma_device_type": 1 00:15:18.980 }, 00:15:18.980 { 00:15:18.980 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:18.980 "dma_device_type": 2 00:15:18.980 } 00:15:18.980 ], 00:15:18.980 "driver_specific": {} 00:15:18.980 } 00:15:18.980 ] 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:18.980 "name": "Existed_Raid", 00:15:18.980 "uuid": "7a9c8f81-9d45-423c-889e-ed603fa66856", 00:15:18.980 "strip_size_kb": 0, 00:15:18.980 "state": "configuring", 00:15:18.980 "raid_level": "raid1", 00:15:18.980 "superblock": true, 00:15:18.980 "num_base_bdevs": 2, 00:15:18.980 "num_base_bdevs_discovered": 1, 00:15:18.980 "num_base_bdevs_operational": 2, 00:15:18.980 "base_bdevs_list": [ 00:15:18.980 { 00:15:18.980 "name": "BaseBdev1", 00:15:18.980 "uuid": "5da35ddd-8ba0-4d99-811d-56f9700ee42d", 00:15:18.980 "is_configured": true, 00:15:18.980 "data_offset": 256, 00:15:18.980 "data_size": 7936 00:15:18.980 }, 00:15:18.980 { 00:15:18.980 "name": "BaseBdev2", 00:15:18.980 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:18.980 "is_configured": false, 00:15:18.980 "data_offset": 0, 00:15:18.980 "data_size": 0 00:15:18.980 } 00:15:18.980 ] 00:15:18.980 }' 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:18.980 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.241 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:19.241 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.241 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.241 [2024-09-29 16:53:40.912214] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:19.241 [2024-09-29 16:53:40.912259] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.558 [2024-09-29 16:53:40.924236] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:19.558 [2024-09-29 16:53:40.926254] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:19.558 [2024-09-29 16:53:40.926366] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:19.558 "name": "Existed_Raid", 00:15:19.558 "uuid": "c4727020-9be4-411e-9608-fc6395d27c9a", 00:15:19.558 "strip_size_kb": 0, 00:15:19.558 "state": "configuring", 00:15:19.558 "raid_level": "raid1", 00:15:19.558 "superblock": true, 00:15:19.558 "num_base_bdevs": 2, 00:15:19.558 "num_base_bdevs_discovered": 1, 00:15:19.558 "num_base_bdevs_operational": 2, 00:15:19.558 "base_bdevs_list": [ 00:15:19.558 { 00:15:19.558 "name": "BaseBdev1", 00:15:19.558 "uuid": "5da35ddd-8ba0-4d99-811d-56f9700ee42d", 00:15:19.558 "is_configured": true, 00:15:19.558 "data_offset": 256, 00:15:19.558 "data_size": 7936 00:15:19.558 }, 00:15:19.558 { 00:15:19.558 "name": "BaseBdev2", 00:15:19.558 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.558 "is_configured": false, 00:15:19.558 "data_offset": 0, 00:15:19.558 "data_size": 0 00:15:19.558 } 00:15:19.558 ] 00:15:19.558 }' 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:19.558 16:53:40 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.844 [2024-09-29 16:53:41.353635] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:19.844 [2024-09-29 16:53:41.354515] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:19.844 [2024-09-29 16:53:41.354754] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:19.844 BaseBdev2 00:15:19.844 [2024-09-29 16:53:41.355816] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.844 [2024-09-29 16:53:41.356338] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:19.844 [2024-09-29 16:53:41.356512] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:19.844 [2024-09-29 16:53:41.356916] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:19.844 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.845 [ 00:15:19.845 { 00:15:19.845 "name": "BaseBdev2", 00:15:19.845 "aliases": [ 00:15:19.845 "68273540-a1d7-4377-b262-75b56d154272" 00:15:19.845 ], 00:15:19.845 "product_name": "Malloc disk", 00:15:19.845 "block_size": 4096, 00:15:19.845 "num_blocks": 8192, 00:15:19.845 "uuid": "68273540-a1d7-4377-b262-75b56d154272", 00:15:19.845 "assigned_rate_limits": { 00:15:19.845 "rw_ios_per_sec": 0, 00:15:19.845 "rw_mbytes_per_sec": 0, 00:15:19.845 "r_mbytes_per_sec": 0, 00:15:19.845 "w_mbytes_per_sec": 0 00:15:19.845 }, 00:15:19.845 "claimed": true, 00:15:19.845 "claim_type": "exclusive_write", 00:15:19.845 "zoned": false, 00:15:19.845 "supported_io_types": { 00:15:19.845 "read": true, 00:15:19.845 "write": true, 00:15:19.845 "unmap": true, 00:15:19.845 "flush": true, 00:15:19.845 "reset": true, 00:15:19.845 "nvme_admin": false, 00:15:19.845 "nvme_io": false, 00:15:19.845 "nvme_io_md": false, 00:15:19.845 "write_zeroes": true, 00:15:19.845 "zcopy": true, 00:15:19.845 "get_zone_info": false, 00:15:19.845 "zone_management": false, 00:15:19.845 "zone_append": false, 00:15:19.845 "compare": false, 00:15:19.845 "compare_and_write": false, 00:15:19.845 "abort": true, 00:15:19.845 "seek_hole": false, 00:15:19.845 "seek_data": false, 00:15:19.845 "copy": true, 00:15:19.845 "nvme_iov_md": false 00:15:19.845 }, 00:15:19.845 "memory_domains": [ 00:15:19.845 { 00:15:19.845 "dma_device_id": "system", 00:15:19.845 "dma_device_type": 1 00:15:19.845 }, 00:15:19.845 { 00:15:19.845 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:19.845 "dma_device_type": 2 00:15:19.845 } 00:15:19.845 ], 00:15:19.845 "driver_specific": {} 00:15:19.845 } 00:15:19.845 ] 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:19.845 "name": "Existed_Raid", 00:15:19.845 "uuid": "c4727020-9be4-411e-9608-fc6395d27c9a", 00:15:19.845 "strip_size_kb": 0, 00:15:19.845 "state": "online", 00:15:19.845 "raid_level": "raid1", 00:15:19.845 "superblock": true, 00:15:19.845 "num_base_bdevs": 2, 00:15:19.845 "num_base_bdevs_discovered": 2, 00:15:19.845 "num_base_bdevs_operational": 2, 00:15:19.845 "base_bdevs_list": [ 00:15:19.845 { 00:15:19.845 "name": "BaseBdev1", 00:15:19.845 "uuid": "5da35ddd-8ba0-4d99-811d-56f9700ee42d", 00:15:19.845 "is_configured": true, 00:15:19.845 "data_offset": 256, 00:15:19.845 "data_size": 7936 00:15:19.845 }, 00:15:19.845 { 00:15:19.845 "name": "BaseBdev2", 00:15:19.845 "uuid": "68273540-a1d7-4377-b262-75b56d154272", 00:15:19.845 "is_configured": true, 00:15:19.845 "data_offset": 256, 00:15:19.845 "data_size": 7936 00:15:19.845 } 00:15:19.845 ] 00:15:19.845 }' 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:19.845 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.415 [2024-09-29 16:53:41.869015] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:20.415 "name": "Existed_Raid", 00:15:20.415 "aliases": [ 00:15:20.415 "c4727020-9be4-411e-9608-fc6395d27c9a" 00:15:20.415 ], 00:15:20.415 "product_name": "Raid Volume", 00:15:20.415 "block_size": 4096, 00:15:20.415 "num_blocks": 7936, 00:15:20.415 "uuid": "c4727020-9be4-411e-9608-fc6395d27c9a", 00:15:20.415 "assigned_rate_limits": { 00:15:20.415 "rw_ios_per_sec": 0, 00:15:20.415 "rw_mbytes_per_sec": 0, 00:15:20.415 "r_mbytes_per_sec": 0, 00:15:20.415 "w_mbytes_per_sec": 0 00:15:20.415 }, 00:15:20.415 "claimed": false, 00:15:20.415 "zoned": false, 00:15:20.415 "supported_io_types": { 00:15:20.415 "read": true, 00:15:20.415 "write": true, 00:15:20.415 "unmap": false, 00:15:20.415 "flush": false, 00:15:20.415 "reset": true, 00:15:20.415 "nvme_admin": false, 00:15:20.415 "nvme_io": false, 00:15:20.415 "nvme_io_md": false, 00:15:20.415 "write_zeroes": true, 00:15:20.415 "zcopy": false, 00:15:20.415 "get_zone_info": false, 00:15:20.415 "zone_management": false, 00:15:20.415 "zone_append": false, 00:15:20.415 "compare": false, 00:15:20.415 "compare_and_write": false, 00:15:20.415 "abort": false, 00:15:20.415 "seek_hole": false, 00:15:20.415 "seek_data": false, 00:15:20.415 "copy": false, 00:15:20.415 "nvme_iov_md": false 00:15:20.415 }, 00:15:20.415 "memory_domains": [ 00:15:20.415 { 00:15:20.415 "dma_device_id": "system", 00:15:20.415 "dma_device_type": 1 00:15:20.415 }, 00:15:20.415 { 00:15:20.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:20.415 "dma_device_type": 2 00:15:20.415 }, 00:15:20.415 { 00:15:20.415 "dma_device_id": "system", 00:15:20.415 "dma_device_type": 1 00:15:20.415 }, 00:15:20.415 { 00:15:20.415 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:20.415 "dma_device_type": 2 00:15:20.415 } 00:15:20.415 ], 00:15:20.415 "driver_specific": { 00:15:20.415 "raid": { 00:15:20.415 "uuid": "c4727020-9be4-411e-9608-fc6395d27c9a", 00:15:20.415 "strip_size_kb": 0, 00:15:20.415 "state": "online", 00:15:20.415 "raid_level": "raid1", 00:15:20.415 "superblock": true, 00:15:20.415 "num_base_bdevs": 2, 00:15:20.415 "num_base_bdevs_discovered": 2, 00:15:20.415 "num_base_bdevs_operational": 2, 00:15:20.415 "base_bdevs_list": [ 00:15:20.415 { 00:15:20.415 "name": "BaseBdev1", 00:15:20.415 "uuid": "5da35ddd-8ba0-4d99-811d-56f9700ee42d", 00:15:20.415 "is_configured": true, 00:15:20.415 "data_offset": 256, 00:15:20.415 "data_size": 7936 00:15:20.415 }, 00:15:20.415 { 00:15:20.415 "name": "BaseBdev2", 00:15:20.415 "uuid": "68273540-a1d7-4377-b262-75b56d154272", 00:15:20.415 "is_configured": true, 00:15:20.415 "data_offset": 256, 00:15:20.415 "data_size": 7936 00:15:20.415 } 00:15:20.415 ] 00:15:20.415 } 00:15:20.415 } 00:15:20.415 }' 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:20.415 BaseBdev2' 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.415 16:53:41 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.415 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.415 [2024-09-29 16:53:42.076449] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:20.675 "name": "Existed_Raid", 00:15:20.675 "uuid": "c4727020-9be4-411e-9608-fc6395d27c9a", 00:15:20.675 "strip_size_kb": 0, 00:15:20.675 "state": "online", 00:15:20.675 "raid_level": "raid1", 00:15:20.675 "superblock": true, 00:15:20.675 "num_base_bdevs": 2, 00:15:20.675 "num_base_bdevs_discovered": 1, 00:15:20.675 "num_base_bdevs_operational": 1, 00:15:20.675 "base_bdevs_list": [ 00:15:20.675 { 00:15:20.675 "name": null, 00:15:20.675 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:20.675 "is_configured": false, 00:15:20.675 "data_offset": 0, 00:15:20.675 "data_size": 7936 00:15:20.675 }, 00:15:20.675 { 00:15:20.675 "name": "BaseBdev2", 00:15:20.675 "uuid": "68273540-a1d7-4377-b262-75b56d154272", 00:15:20.675 "is_configured": true, 00:15:20.675 "data_offset": 256, 00:15:20.675 "data_size": 7936 00:15:20.675 } 00:15:20.675 ] 00:15:20.675 }' 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:20.675 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.935 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:20.935 [2024-09-29 16:53:42.599371] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:20.935 [2024-09-29 16:53:42.599510] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:21.194 [2024-09-29 16:53:42.611539] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:21.194 [2024-09-29 16:53:42.611593] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:21.194 [2024-09-29 16:53:42.611605] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 95950 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 95950 ']' 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 95950 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 95950 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:21.194 killing process with pid 95950 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 95950' 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@969 -- # kill 95950 00:15:21.194 [2024-09-29 16:53:42.702873] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:21.194 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@974 -- # wait 95950 00:15:21.194 [2024-09-29 16:53:42.703903] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:21.454 ************************************ 00:15:21.454 END TEST raid_state_function_test_sb_4k 00:15:21.454 ************************************ 00:15:21.454 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:15:21.454 00:15:21.454 real 0m3.991s 00:15:21.454 user 0m6.224s 00:15:21.454 sys 0m0.885s 00:15:21.454 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:21.454 16:53:42 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.454 16:53:43 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:15:21.454 16:53:43 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:15:21.454 16:53:43 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:21.454 16:53:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:21.454 ************************************ 00:15:21.454 START TEST raid_superblock_test_4k 00:15:21.454 ************************************ 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=96191 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 96191 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@831 -- # '[' -z 96191 ']' 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:21.454 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:21.454 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:21.454 [2024-09-29 16:53:43.120311] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:15:21.454 [2024-09-29 16:53:43.120435] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96191 ] 00:15:21.714 [2024-09-29 16:53:43.267132] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:21.714 [2024-09-29 16:53:43.313110] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:21.714 [2024-09-29 16:53:43.355876] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:21.714 [2024-09-29 16:53:43.355922] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # return 0 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.284 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.544 malloc1 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.544 [2024-09-29 16:53:43.966319] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:22.544 [2024-09-29 16:53:43.966444] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:22.544 [2024-09-29 16:53:43.966481] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:22.544 [2024-09-29 16:53:43.966525] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:22.544 [2024-09-29 16:53:43.968792] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:22.544 [2024-09-29 16:53:43.968887] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:22.544 pt1 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.544 16:53:43 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.544 malloc2 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.544 [2024-09-29 16:53:44.018334] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:22.544 [2024-09-29 16:53:44.018522] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:22.544 [2024-09-29 16:53:44.018624] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:22.544 [2024-09-29 16:53:44.018748] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:22.544 [2024-09-29 16:53:44.023419] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:22.544 [2024-09-29 16:53:44.023471] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:22.544 pt2 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.544 [2024-09-29 16:53:44.031756] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:22.544 [2024-09-29 16:53:44.034123] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:22.544 [2024-09-29 16:53:44.034342] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:22.544 [2024-09-29 16:53:44.034366] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:22.544 [2024-09-29 16:53:44.034692] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:22.544 [2024-09-29 16:53:44.034861] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:22.544 [2024-09-29 16:53:44.034885] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:22.544 [2024-09-29 16:53:44.035050] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:22.544 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:22.545 "name": "raid_bdev1", 00:15:22.545 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:22.545 "strip_size_kb": 0, 00:15:22.545 "state": "online", 00:15:22.545 "raid_level": "raid1", 00:15:22.545 "superblock": true, 00:15:22.545 "num_base_bdevs": 2, 00:15:22.545 "num_base_bdevs_discovered": 2, 00:15:22.545 "num_base_bdevs_operational": 2, 00:15:22.545 "base_bdevs_list": [ 00:15:22.545 { 00:15:22.545 "name": "pt1", 00:15:22.545 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:22.545 "is_configured": true, 00:15:22.545 "data_offset": 256, 00:15:22.545 "data_size": 7936 00:15:22.545 }, 00:15:22.545 { 00:15:22.545 "name": "pt2", 00:15:22.545 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:22.545 "is_configured": true, 00:15:22.545 "data_offset": 256, 00:15:22.545 "data_size": 7936 00:15:22.545 } 00:15:22.545 ] 00:15:22.545 }' 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:22.545 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.114 [2024-09-29 16:53:44.535232] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:23.114 "name": "raid_bdev1", 00:15:23.114 "aliases": [ 00:15:23.114 "a585742b-34bd-4f5c-994c-e1cf02262a93" 00:15:23.114 ], 00:15:23.114 "product_name": "Raid Volume", 00:15:23.114 "block_size": 4096, 00:15:23.114 "num_blocks": 7936, 00:15:23.114 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:23.114 "assigned_rate_limits": { 00:15:23.114 "rw_ios_per_sec": 0, 00:15:23.114 "rw_mbytes_per_sec": 0, 00:15:23.114 "r_mbytes_per_sec": 0, 00:15:23.114 "w_mbytes_per_sec": 0 00:15:23.114 }, 00:15:23.114 "claimed": false, 00:15:23.114 "zoned": false, 00:15:23.114 "supported_io_types": { 00:15:23.114 "read": true, 00:15:23.114 "write": true, 00:15:23.114 "unmap": false, 00:15:23.114 "flush": false, 00:15:23.114 "reset": true, 00:15:23.114 "nvme_admin": false, 00:15:23.114 "nvme_io": false, 00:15:23.114 "nvme_io_md": false, 00:15:23.114 "write_zeroes": true, 00:15:23.114 "zcopy": false, 00:15:23.114 "get_zone_info": false, 00:15:23.114 "zone_management": false, 00:15:23.114 "zone_append": false, 00:15:23.114 "compare": false, 00:15:23.114 "compare_and_write": false, 00:15:23.114 "abort": false, 00:15:23.114 "seek_hole": false, 00:15:23.114 "seek_data": false, 00:15:23.114 "copy": false, 00:15:23.114 "nvme_iov_md": false 00:15:23.114 }, 00:15:23.114 "memory_domains": [ 00:15:23.114 { 00:15:23.114 "dma_device_id": "system", 00:15:23.114 "dma_device_type": 1 00:15:23.114 }, 00:15:23.114 { 00:15:23.114 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:23.114 "dma_device_type": 2 00:15:23.114 }, 00:15:23.114 { 00:15:23.114 "dma_device_id": "system", 00:15:23.114 "dma_device_type": 1 00:15:23.114 }, 00:15:23.114 { 00:15:23.114 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:23.114 "dma_device_type": 2 00:15:23.114 } 00:15:23.114 ], 00:15:23.114 "driver_specific": { 00:15:23.114 "raid": { 00:15:23.114 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:23.114 "strip_size_kb": 0, 00:15:23.114 "state": "online", 00:15:23.114 "raid_level": "raid1", 00:15:23.114 "superblock": true, 00:15:23.114 "num_base_bdevs": 2, 00:15:23.114 "num_base_bdevs_discovered": 2, 00:15:23.114 "num_base_bdevs_operational": 2, 00:15:23.114 "base_bdevs_list": [ 00:15:23.114 { 00:15:23.114 "name": "pt1", 00:15:23.114 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:23.114 "is_configured": true, 00:15:23.114 "data_offset": 256, 00:15:23.114 "data_size": 7936 00:15:23.114 }, 00:15:23.114 { 00:15:23.114 "name": "pt2", 00:15:23.114 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:23.114 "is_configured": true, 00:15:23.114 "data_offset": 256, 00:15:23.114 "data_size": 7936 00:15:23.114 } 00:15:23.114 ] 00:15:23.114 } 00:15:23.114 } 00:15:23.114 }' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:23.114 pt2' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.114 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.114 [2024-09-29 16:53:44.766800] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a585742b-34bd-4f5c-994c-e1cf02262a93 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z a585742b-34bd-4f5c-994c-e1cf02262a93 ']' 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 [2024-09-29 16:53:44.810504] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:23.375 [2024-09-29 16:53:44.810573] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:23.375 [2024-09-29 16:53:44.810652] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:23.375 [2024-09-29 16:53:44.810752] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:23.375 [2024-09-29 16:53:44.810807] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@650 -- # local es=0 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 [2024-09-29 16:53:44.950278] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:23.375 [2024-09-29 16:53:44.952104] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:23.375 [2024-09-29 16:53:44.952193] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:23.375 [2024-09-29 16:53:44.952318] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:23.375 [2024-09-29 16:53:44.952387] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:23.375 [2024-09-29 16:53:44.952433] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:23.375 request: 00:15:23.375 { 00:15:23.375 "name": "raid_bdev1", 00:15:23.375 "raid_level": "raid1", 00:15:23.375 "base_bdevs": [ 00:15:23.375 "malloc1", 00:15:23.375 "malloc2" 00:15:23.375 ], 00:15:23.375 "superblock": false, 00:15:23.375 "method": "bdev_raid_create", 00:15:23.375 "req_id": 1 00:15:23.375 } 00:15:23.375 Got JSON-RPC error response 00:15:23.375 response: 00:15:23.375 { 00:15:23.375 "code": -17, 00:15:23.375 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:23.375 } 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # es=1 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 16:53:44 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.375 [2024-09-29 16:53:45.018136] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:23.375 [2024-09-29 16:53:45.018224] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:23.375 [2024-09-29 16:53:45.018261] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:23.375 [2024-09-29 16:53:45.018286] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:23.375 [2024-09-29 16:53:45.020366] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:23.375 [2024-09-29 16:53:45.020462] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:23.375 [2024-09-29 16:53:45.020541] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:23.375 [2024-09-29 16:53:45.020592] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:23.375 pt1 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:23.375 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.376 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.635 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.635 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:23.635 "name": "raid_bdev1", 00:15:23.635 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:23.635 "strip_size_kb": 0, 00:15:23.635 "state": "configuring", 00:15:23.635 "raid_level": "raid1", 00:15:23.635 "superblock": true, 00:15:23.635 "num_base_bdevs": 2, 00:15:23.635 "num_base_bdevs_discovered": 1, 00:15:23.635 "num_base_bdevs_operational": 2, 00:15:23.635 "base_bdevs_list": [ 00:15:23.635 { 00:15:23.635 "name": "pt1", 00:15:23.635 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:23.635 "is_configured": true, 00:15:23.635 "data_offset": 256, 00:15:23.635 "data_size": 7936 00:15:23.635 }, 00:15:23.635 { 00:15:23.635 "name": null, 00:15:23.635 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:23.635 "is_configured": false, 00:15:23.635 "data_offset": 256, 00:15:23.635 "data_size": 7936 00:15:23.635 } 00:15:23.635 ] 00:15:23.635 }' 00:15:23.635 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:23.635 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.895 [2024-09-29 16:53:45.473369] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:23.895 [2024-09-29 16:53:45.473451] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:23.895 [2024-09-29 16:53:45.473483] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:23.895 [2024-09-29 16:53:45.473509] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:23.895 [2024-09-29 16:53:45.473887] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:23.895 [2024-09-29 16:53:45.473942] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:23.895 [2024-09-29 16:53:45.474034] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:23.895 [2024-09-29 16:53:45.474079] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:23.895 [2024-09-29 16:53:45.474196] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:23.895 [2024-09-29 16:53:45.474233] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:23.895 [2024-09-29 16:53:45.474496] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:23.895 [2024-09-29 16:53:45.474646] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:23.895 [2024-09-29 16:53:45.474690] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:23.895 [2024-09-29 16:53:45.474855] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:23.895 pt2 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:23.895 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:23.895 "name": "raid_bdev1", 00:15:23.895 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:23.895 "strip_size_kb": 0, 00:15:23.895 "state": "online", 00:15:23.895 "raid_level": "raid1", 00:15:23.895 "superblock": true, 00:15:23.895 "num_base_bdevs": 2, 00:15:23.895 "num_base_bdevs_discovered": 2, 00:15:23.895 "num_base_bdevs_operational": 2, 00:15:23.895 "base_bdevs_list": [ 00:15:23.895 { 00:15:23.895 "name": "pt1", 00:15:23.895 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:23.895 "is_configured": true, 00:15:23.895 "data_offset": 256, 00:15:23.895 "data_size": 7936 00:15:23.895 }, 00:15:23.895 { 00:15:23.895 "name": "pt2", 00:15:23.895 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:23.895 "is_configured": true, 00:15:23.895 "data_offset": 256, 00:15:23.895 "data_size": 7936 00:15:23.895 } 00:15:23.895 ] 00:15:23.895 }' 00:15:23.896 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:23.896 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.464 [2024-09-29 16:53:45.908858] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.464 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:24.464 "name": "raid_bdev1", 00:15:24.464 "aliases": [ 00:15:24.464 "a585742b-34bd-4f5c-994c-e1cf02262a93" 00:15:24.464 ], 00:15:24.464 "product_name": "Raid Volume", 00:15:24.464 "block_size": 4096, 00:15:24.464 "num_blocks": 7936, 00:15:24.464 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:24.464 "assigned_rate_limits": { 00:15:24.464 "rw_ios_per_sec": 0, 00:15:24.464 "rw_mbytes_per_sec": 0, 00:15:24.464 "r_mbytes_per_sec": 0, 00:15:24.464 "w_mbytes_per_sec": 0 00:15:24.464 }, 00:15:24.464 "claimed": false, 00:15:24.464 "zoned": false, 00:15:24.464 "supported_io_types": { 00:15:24.464 "read": true, 00:15:24.464 "write": true, 00:15:24.464 "unmap": false, 00:15:24.464 "flush": false, 00:15:24.464 "reset": true, 00:15:24.464 "nvme_admin": false, 00:15:24.464 "nvme_io": false, 00:15:24.464 "nvme_io_md": false, 00:15:24.464 "write_zeroes": true, 00:15:24.464 "zcopy": false, 00:15:24.464 "get_zone_info": false, 00:15:24.464 "zone_management": false, 00:15:24.464 "zone_append": false, 00:15:24.464 "compare": false, 00:15:24.464 "compare_and_write": false, 00:15:24.465 "abort": false, 00:15:24.465 "seek_hole": false, 00:15:24.465 "seek_data": false, 00:15:24.465 "copy": false, 00:15:24.465 "nvme_iov_md": false 00:15:24.465 }, 00:15:24.465 "memory_domains": [ 00:15:24.465 { 00:15:24.465 "dma_device_id": "system", 00:15:24.465 "dma_device_type": 1 00:15:24.465 }, 00:15:24.465 { 00:15:24.465 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:24.465 "dma_device_type": 2 00:15:24.465 }, 00:15:24.465 { 00:15:24.465 "dma_device_id": "system", 00:15:24.465 "dma_device_type": 1 00:15:24.465 }, 00:15:24.465 { 00:15:24.465 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:24.465 "dma_device_type": 2 00:15:24.465 } 00:15:24.465 ], 00:15:24.465 "driver_specific": { 00:15:24.465 "raid": { 00:15:24.465 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:24.465 "strip_size_kb": 0, 00:15:24.465 "state": "online", 00:15:24.465 "raid_level": "raid1", 00:15:24.465 "superblock": true, 00:15:24.465 "num_base_bdevs": 2, 00:15:24.465 "num_base_bdevs_discovered": 2, 00:15:24.465 "num_base_bdevs_operational": 2, 00:15:24.465 "base_bdevs_list": [ 00:15:24.465 { 00:15:24.465 "name": "pt1", 00:15:24.465 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:24.465 "is_configured": true, 00:15:24.465 "data_offset": 256, 00:15:24.465 "data_size": 7936 00:15:24.465 }, 00:15:24.465 { 00:15:24.465 "name": "pt2", 00:15:24.465 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:24.465 "is_configured": true, 00:15:24.465 "data_offset": 256, 00:15:24.465 "data_size": 7936 00:15:24.465 } 00:15:24.465 ] 00:15:24.465 } 00:15:24.465 } 00:15:24.465 }' 00:15:24.465 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:24.465 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:24.465 pt2' 00:15:24.465 16:53:45 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:24.465 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.725 [2024-09-29 16:53:46.140466] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' a585742b-34bd-4f5c-994c-e1cf02262a93 '!=' a585742b-34bd-4f5c-994c-e1cf02262a93 ']' 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.725 [2024-09-29 16:53:46.188200] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.725 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.726 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.726 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:24.726 "name": "raid_bdev1", 00:15:24.726 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:24.726 "strip_size_kb": 0, 00:15:24.726 "state": "online", 00:15:24.726 "raid_level": "raid1", 00:15:24.726 "superblock": true, 00:15:24.726 "num_base_bdevs": 2, 00:15:24.726 "num_base_bdevs_discovered": 1, 00:15:24.726 "num_base_bdevs_operational": 1, 00:15:24.726 "base_bdevs_list": [ 00:15:24.726 { 00:15:24.726 "name": null, 00:15:24.726 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:24.726 "is_configured": false, 00:15:24.726 "data_offset": 0, 00:15:24.726 "data_size": 7936 00:15:24.726 }, 00:15:24.726 { 00:15:24.726 "name": "pt2", 00:15:24.726 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:24.726 "is_configured": true, 00:15:24.726 "data_offset": 256, 00:15:24.726 "data_size": 7936 00:15:24.726 } 00:15:24.726 ] 00:15:24.726 }' 00:15:24.726 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:24.726 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.985 [2024-09-29 16:53:46.631387] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:24.985 [2024-09-29 16:53:46.631409] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:24.985 [2024-09-29 16:53:46.631462] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:24.985 [2024-09-29 16:53:46.631496] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:24.985 [2024-09-29 16:53:46.631504] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:24.985 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.246 [2024-09-29 16:53:46.707325] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:25.246 [2024-09-29 16:53:46.707422] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:25.246 [2024-09-29 16:53:46.707457] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:15:25.246 [2024-09-29 16:53:46.707482] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:25.246 [2024-09-29 16:53:46.709516] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:25.246 [2024-09-29 16:53:46.709589] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:25.246 [2024-09-29 16:53:46.709672] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:25.246 [2024-09-29 16:53:46.709745] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:25.246 [2024-09-29 16:53:46.709854] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:25.246 [2024-09-29 16:53:46.709891] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:25.246 [2024-09-29 16:53:46.710149] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:25.246 [2024-09-29 16:53:46.710290] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:25.246 [2024-09-29 16:53:46.710331] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:25.246 [2024-09-29 16:53:46.710486] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:25.246 pt2 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:25.246 "name": "raid_bdev1", 00:15:25.246 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:25.246 "strip_size_kb": 0, 00:15:25.246 "state": "online", 00:15:25.246 "raid_level": "raid1", 00:15:25.246 "superblock": true, 00:15:25.246 "num_base_bdevs": 2, 00:15:25.246 "num_base_bdevs_discovered": 1, 00:15:25.246 "num_base_bdevs_operational": 1, 00:15:25.246 "base_bdevs_list": [ 00:15:25.246 { 00:15:25.246 "name": null, 00:15:25.246 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:25.246 "is_configured": false, 00:15:25.246 "data_offset": 256, 00:15:25.246 "data_size": 7936 00:15:25.246 }, 00:15:25.246 { 00:15:25.246 "name": "pt2", 00:15:25.246 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:25.246 "is_configured": true, 00:15:25.246 "data_offset": 256, 00:15:25.246 "data_size": 7936 00:15:25.246 } 00:15:25.246 ] 00:15:25.246 }' 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:25.246 16:53:46 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.506 [2024-09-29 16:53:47.142792] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:25.506 [2024-09-29 16:53:47.142813] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:25.506 [2024-09-29 16:53:47.142858] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:25.506 [2024-09-29 16:53:47.142889] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:25.506 [2024-09-29 16:53:47.142899] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.506 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.766 [2024-09-29 16:53:47.202714] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:25.766 [2024-09-29 16:53:47.202810] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:25.766 [2024-09-29 16:53:47.202840] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:15:25.766 [2024-09-29 16:53:47.202868] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:25.766 [2024-09-29 16:53:47.205009] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:25.766 [2024-09-29 16:53:47.205081] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:25.766 [2024-09-29 16:53:47.205154] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:25.766 [2024-09-29 16:53:47.205252] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:25.766 [2024-09-29 16:53:47.205383] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:25.766 [2024-09-29 16:53:47.205436] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:25.766 [2024-09-29 16:53:47.205498] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:25.766 [2024-09-29 16:53:47.205540] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:25.766 [2024-09-29 16:53:47.205604] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:25.766 [2024-09-29 16:53:47.205614] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:25.766 [2024-09-29 16:53:47.205824] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:25.766 [2024-09-29 16:53:47.205932] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:25.766 [2024-09-29 16:53:47.205941] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:25.766 [2024-09-29 16:53:47.206043] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:25.766 pt1 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:25.766 "name": "raid_bdev1", 00:15:25.766 "uuid": "a585742b-34bd-4f5c-994c-e1cf02262a93", 00:15:25.766 "strip_size_kb": 0, 00:15:25.766 "state": "online", 00:15:25.766 "raid_level": "raid1", 00:15:25.766 "superblock": true, 00:15:25.766 "num_base_bdevs": 2, 00:15:25.766 "num_base_bdevs_discovered": 1, 00:15:25.766 "num_base_bdevs_operational": 1, 00:15:25.766 "base_bdevs_list": [ 00:15:25.766 { 00:15:25.766 "name": null, 00:15:25.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:25.766 "is_configured": false, 00:15:25.766 "data_offset": 256, 00:15:25.766 "data_size": 7936 00:15:25.766 }, 00:15:25.766 { 00:15:25.766 "name": "pt2", 00:15:25.766 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:25.766 "is_configured": true, 00:15:25.766 "data_offset": 256, 00:15:25.766 "data_size": 7936 00:15:25.766 } 00:15:25.766 ] 00:15:25.766 }' 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:25.766 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.026 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:26.026 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:26.026 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.026 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.026 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.286 [2024-09-29 16:53:47.730014] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' a585742b-34bd-4f5c-994c-e1cf02262a93 '!=' a585742b-34bd-4f5c-994c-e1cf02262a93 ']' 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 96191 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@950 -- # '[' -z 96191 ']' 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # kill -0 96191 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # uname 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 96191 00:15:26.286 killing process with pid 96191 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 96191' 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@969 -- # kill 96191 00:15:26.286 [2024-09-29 16:53:47.799876] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:26.286 [2024-09-29 16:53:47.799927] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:26.286 [2024-09-29 16:53:47.799961] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:26.286 [2024-09-29 16:53:47.799968] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:26.286 16:53:47 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@974 -- # wait 96191 00:15:26.286 [2024-09-29 16:53:47.823210] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:26.546 16:53:48 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:15:26.546 00:15:26.546 real 0m5.039s 00:15:26.546 user 0m8.200s 00:15:26.546 sys 0m1.099s 00:15:26.546 16:53:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:26.546 ************************************ 00:15:26.546 END TEST raid_superblock_test_4k 00:15:26.546 ************************************ 00:15:26.546 16:53:48 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.546 16:53:48 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:15:26.546 16:53:48 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:15:26.546 16:53:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:15:26.546 16:53:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:26.546 16:53:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:26.546 ************************************ 00:15:26.546 START TEST raid_rebuild_test_sb_4k 00:15:26.546 ************************************ 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:26.546 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=96503 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 96503 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 96503 ']' 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:26.547 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:26.547 16:53:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:26.807 [2024-09-29 16:53:48.249788] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:15:26.807 [2024-09-29 16:53:48.250009] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:15:26.807 Zero copy mechanism will not be used. 00:15:26.807 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96503 ] 00:15:26.807 [2024-09-29 16:53:48.396258] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:26.807 [2024-09-29 16:53:48.441660] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:27.067 [2024-09-29 16:53:48.485446] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:27.067 [2024-09-29 16:53:48.485536] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 BaseBdev1_malloc 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 [2024-09-29 16:53:49.091759] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:27.636 [2024-09-29 16:53:49.091818] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:27.636 [2024-09-29 16:53:49.091841] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:27.636 [2024-09-29 16:53:49.091854] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:27.636 [2024-09-29 16:53:49.093944] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:27.636 [2024-09-29 16:53:49.093978] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:27.636 BaseBdev1 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 BaseBdev2_malloc 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 [2024-09-29 16:53:49.133445] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:27.636 [2024-09-29 16:53:49.133535] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:27.636 [2024-09-29 16:53:49.133575] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:27.636 [2024-09-29 16:53:49.133595] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:27.636 [2024-09-29 16:53:49.138129] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:27.636 [2024-09-29 16:53:49.138197] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:27.636 BaseBdev2 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 spare_malloc 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 spare_delay 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 [2024-09-29 16:53:49.176351] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:27.636 [2024-09-29 16:53:49.176444] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:27.636 [2024-09-29 16:53:49.176470] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:27.636 [2024-09-29 16:53:49.176478] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:27.636 [2024-09-29 16:53:49.178573] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:27.636 [2024-09-29 16:53:49.178611] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:27.636 spare 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 [2024-09-29 16:53:49.188382] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:27.636 [2024-09-29 16:53:49.190134] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:27.636 [2024-09-29 16:53:49.190290] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:27.636 [2024-09-29 16:53:49.190307] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:27.636 [2024-09-29 16:53:49.190558] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:27.636 [2024-09-29 16:53:49.190681] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:27.636 [2024-09-29 16:53:49.190692] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:27.636 [2024-09-29 16:53:49.190826] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.636 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:27.636 "name": "raid_bdev1", 00:15:27.636 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:27.636 "strip_size_kb": 0, 00:15:27.636 "state": "online", 00:15:27.636 "raid_level": "raid1", 00:15:27.636 "superblock": true, 00:15:27.636 "num_base_bdevs": 2, 00:15:27.636 "num_base_bdevs_discovered": 2, 00:15:27.636 "num_base_bdevs_operational": 2, 00:15:27.636 "base_bdevs_list": [ 00:15:27.636 { 00:15:27.636 "name": "BaseBdev1", 00:15:27.636 "uuid": "5e465916-7341-5b04-8161-25cdee2a6a4a", 00:15:27.636 "is_configured": true, 00:15:27.636 "data_offset": 256, 00:15:27.636 "data_size": 7936 00:15:27.636 }, 00:15:27.636 { 00:15:27.636 "name": "BaseBdev2", 00:15:27.637 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:27.637 "is_configured": true, 00:15:27.637 "data_offset": 256, 00:15:27.637 "data_size": 7936 00:15:27.637 } 00:15:27.637 ] 00:15:27.637 }' 00:15:27.637 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:27.637 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:28.205 [2024-09-29 16:53:49.659781] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:28.205 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:28.464 [2024-09-29 16:53:49.907244] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:28.464 /dev/nbd0 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:28.464 1+0 records in 00:15:28.464 1+0 records out 00:15:28.464 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000434106 s, 9.4 MB/s 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:28.464 16:53:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:29.034 7936+0 records in 00:15:29.034 7936+0 records out 00:15:29.034 32505856 bytes (33 MB, 31 MiB) copied, 0.609618 s, 53.3 MB/s 00:15:29.034 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:29.034 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:29.034 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:29.034 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:29.034 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:29.034 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:29.034 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:29.294 [2024-09-29 16:53:50.841637] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.294 [2024-09-29 16:53:50.872740] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:29.294 "name": "raid_bdev1", 00:15:29.294 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:29.294 "strip_size_kb": 0, 00:15:29.294 "state": "online", 00:15:29.294 "raid_level": "raid1", 00:15:29.294 "superblock": true, 00:15:29.294 "num_base_bdevs": 2, 00:15:29.294 "num_base_bdevs_discovered": 1, 00:15:29.294 "num_base_bdevs_operational": 1, 00:15:29.294 "base_bdevs_list": [ 00:15:29.294 { 00:15:29.294 "name": null, 00:15:29.294 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:29.294 "is_configured": false, 00:15:29.294 "data_offset": 0, 00:15:29.294 "data_size": 7936 00:15:29.294 }, 00:15:29.294 { 00:15:29.294 "name": "BaseBdev2", 00:15:29.294 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:29.294 "is_configured": true, 00:15:29.294 "data_offset": 256, 00:15:29.294 "data_size": 7936 00:15:29.294 } 00:15:29.294 ] 00:15:29.294 }' 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:29.294 16:53:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.864 16:53:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:29.864 16:53:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.864 16:53:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:29.864 [2024-09-29 16:53:51.335971] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:29.864 [2024-09-29 16:53:51.340169] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:15:29.864 16:53:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.864 16:53:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:29.864 [2024-09-29 16:53:51.342162] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:30.804 "name": "raid_bdev1", 00:15:30.804 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:30.804 "strip_size_kb": 0, 00:15:30.804 "state": "online", 00:15:30.804 "raid_level": "raid1", 00:15:30.804 "superblock": true, 00:15:30.804 "num_base_bdevs": 2, 00:15:30.804 "num_base_bdevs_discovered": 2, 00:15:30.804 "num_base_bdevs_operational": 2, 00:15:30.804 "process": { 00:15:30.804 "type": "rebuild", 00:15:30.804 "target": "spare", 00:15:30.804 "progress": { 00:15:30.804 "blocks": 2560, 00:15:30.804 "percent": 32 00:15:30.804 } 00:15:30.804 }, 00:15:30.804 "base_bdevs_list": [ 00:15:30.804 { 00:15:30.804 "name": "spare", 00:15:30.804 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:30.804 "is_configured": true, 00:15:30.804 "data_offset": 256, 00:15:30.804 "data_size": 7936 00:15:30.804 }, 00:15:30.804 { 00:15:30.804 "name": "BaseBdev2", 00:15:30.804 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:30.804 "is_configured": true, 00:15:30.804 "data_offset": 256, 00:15:30.804 "data_size": 7936 00:15:30.804 } 00:15:30.804 ] 00:15:30.804 }' 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:30.804 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.064 [2024-09-29 16:53:52.483129] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.064 [2024-09-29 16:53:52.546759] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:31.064 [2024-09-29 16:53:52.546807] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:31.064 [2024-09-29 16:53:52.546825] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:31.064 [2024-09-29 16:53:52.546832] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.064 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.065 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.065 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.065 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:31.065 "name": "raid_bdev1", 00:15:31.065 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:31.065 "strip_size_kb": 0, 00:15:31.065 "state": "online", 00:15:31.065 "raid_level": "raid1", 00:15:31.065 "superblock": true, 00:15:31.065 "num_base_bdevs": 2, 00:15:31.065 "num_base_bdevs_discovered": 1, 00:15:31.065 "num_base_bdevs_operational": 1, 00:15:31.065 "base_bdevs_list": [ 00:15:31.065 { 00:15:31.065 "name": null, 00:15:31.065 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.065 "is_configured": false, 00:15:31.065 "data_offset": 0, 00:15:31.065 "data_size": 7936 00:15:31.065 }, 00:15:31.065 { 00:15:31.065 "name": "BaseBdev2", 00:15:31.065 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:31.065 "is_configured": true, 00:15:31.065 "data_offset": 256, 00:15:31.065 "data_size": 7936 00:15:31.065 } 00:15:31.065 ] 00:15:31.065 }' 00:15:31.065 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:31.065 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.324 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:31.324 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:31.324 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:31.324 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:31.324 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:31.584 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:31.584 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.584 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.584 16:53:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:31.584 "name": "raid_bdev1", 00:15:31.584 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:31.584 "strip_size_kb": 0, 00:15:31.584 "state": "online", 00:15:31.584 "raid_level": "raid1", 00:15:31.584 "superblock": true, 00:15:31.584 "num_base_bdevs": 2, 00:15:31.584 "num_base_bdevs_discovered": 1, 00:15:31.584 "num_base_bdevs_operational": 1, 00:15:31.584 "base_bdevs_list": [ 00:15:31.584 { 00:15:31.584 "name": null, 00:15:31.584 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:31.584 "is_configured": false, 00:15:31.584 "data_offset": 0, 00:15:31.584 "data_size": 7936 00:15:31.584 }, 00:15:31.584 { 00:15:31.584 "name": "BaseBdev2", 00:15:31.584 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:31.584 "is_configured": true, 00:15:31.584 "data_offset": 256, 00:15:31.584 "data_size": 7936 00:15:31.584 } 00:15:31.584 ] 00:15:31.584 }' 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:31.584 [2024-09-29 16:53:53.117975] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:31.584 [2024-09-29 16:53:53.121342] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:15:31.584 [2024-09-29 16:53:53.123370] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:31.584 16:53:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:32.524 "name": "raid_bdev1", 00:15:32.524 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:32.524 "strip_size_kb": 0, 00:15:32.524 "state": "online", 00:15:32.524 "raid_level": "raid1", 00:15:32.524 "superblock": true, 00:15:32.524 "num_base_bdevs": 2, 00:15:32.524 "num_base_bdevs_discovered": 2, 00:15:32.524 "num_base_bdevs_operational": 2, 00:15:32.524 "process": { 00:15:32.524 "type": "rebuild", 00:15:32.524 "target": "spare", 00:15:32.524 "progress": { 00:15:32.524 "blocks": 2560, 00:15:32.524 "percent": 32 00:15:32.524 } 00:15:32.524 }, 00:15:32.524 "base_bdevs_list": [ 00:15:32.524 { 00:15:32.524 "name": "spare", 00:15:32.524 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:32.524 "is_configured": true, 00:15:32.524 "data_offset": 256, 00:15:32.524 "data_size": 7936 00:15:32.524 }, 00:15:32.524 { 00:15:32.524 "name": "BaseBdev2", 00:15:32.524 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:32.524 "is_configured": true, 00:15:32.524 "data_offset": 256, 00:15:32.524 "data_size": 7936 00:15:32.524 } 00:15:32.524 ] 00:15:32.524 }' 00:15:32.524 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:32.785 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=559 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:32.785 "name": "raid_bdev1", 00:15:32.785 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:32.785 "strip_size_kb": 0, 00:15:32.785 "state": "online", 00:15:32.785 "raid_level": "raid1", 00:15:32.785 "superblock": true, 00:15:32.785 "num_base_bdevs": 2, 00:15:32.785 "num_base_bdevs_discovered": 2, 00:15:32.785 "num_base_bdevs_operational": 2, 00:15:32.785 "process": { 00:15:32.785 "type": "rebuild", 00:15:32.785 "target": "spare", 00:15:32.785 "progress": { 00:15:32.785 "blocks": 2816, 00:15:32.785 "percent": 35 00:15:32.785 } 00:15:32.785 }, 00:15:32.785 "base_bdevs_list": [ 00:15:32.785 { 00:15:32.785 "name": "spare", 00:15:32.785 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:32.785 "is_configured": true, 00:15:32.785 "data_offset": 256, 00:15:32.785 "data_size": 7936 00:15:32.785 }, 00:15:32.785 { 00:15:32.785 "name": "BaseBdev2", 00:15:32.785 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:32.785 "is_configured": true, 00:15:32.785 "data_offset": 256, 00:15:32.785 "data_size": 7936 00:15:32.785 } 00:15:32.785 ] 00:15:32.785 }' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:32.785 16:53:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:34.167 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:34.167 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:34.168 "name": "raid_bdev1", 00:15:34.168 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:34.168 "strip_size_kb": 0, 00:15:34.168 "state": "online", 00:15:34.168 "raid_level": "raid1", 00:15:34.168 "superblock": true, 00:15:34.168 "num_base_bdevs": 2, 00:15:34.168 "num_base_bdevs_discovered": 2, 00:15:34.168 "num_base_bdevs_operational": 2, 00:15:34.168 "process": { 00:15:34.168 "type": "rebuild", 00:15:34.168 "target": "spare", 00:15:34.168 "progress": { 00:15:34.168 "blocks": 5632, 00:15:34.168 "percent": 70 00:15:34.168 } 00:15:34.168 }, 00:15:34.168 "base_bdevs_list": [ 00:15:34.168 { 00:15:34.168 "name": "spare", 00:15:34.168 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:34.168 "is_configured": true, 00:15:34.168 "data_offset": 256, 00:15:34.168 "data_size": 7936 00:15:34.168 }, 00:15:34.168 { 00:15:34.168 "name": "BaseBdev2", 00:15:34.168 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:34.168 "is_configured": true, 00:15:34.168 "data_offset": 256, 00:15:34.168 "data_size": 7936 00:15:34.168 } 00:15:34.168 ] 00:15:34.168 }' 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:34.168 16:53:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:34.737 [2024-09-29 16:53:56.233612] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:34.737 [2024-09-29 16:53:56.233690] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:34.737 [2024-09-29 16:53:56.233802] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:34.998 "name": "raid_bdev1", 00:15:34.998 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:34.998 "strip_size_kb": 0, 00:15:34.998 "state": "online", 00:15:34.998 "raid_level": "raid1", 00:15:34.998 "superblock": true, 00:15:34.998 "num_base_bdevs": 2, 00:15:34.998 "num_base_bdevs_discovered": 2, 00:15:34.998 "num_base_bdevs_operational": 2, 00:15:34.998 "base_bdevs_list": [ 00:15:34.998 { 00:15:34.998 "name": "spare", 00:15:34.998 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:34.998 "is_configured": true, 00:15:34.998 "data_offset": 256, 00:15:34.998 "data_size": 7936 00:15:34.998 }, 00:15:34.998 { 00:15:34.998 "name": "BaseBdev2", 00:15:34.998 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:34.998 "is_configured": true, 00:15:34.998 "data_offset": 256, 00:15:34.998 "data_size": 7936 00:15:34.998 } 00:15:34.998 ] 00:15:34.998 }' 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:34.998 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:35.258 "name": "raid_bdev1", 00:15:35.258 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:35.258 "strip_size_kb": 0, 00:15:35.258 "state": "online", 00:15:35.258 "raid_level": "raid1", 00:15:35.258 "superblock": true, 00:15:35.258 "num_base_bdevs": 2, 00:15:35.258 "num_base_bdevs_discovered": 2, 00:15:35.258 "num_base_bdevs_operational": 2, 00:15:35.258 "base_bdevs_list": [ 00:15:35.258 { 00:15:35.258 "name": "spare", 00:15:35.258 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:35.258 "is_configured": true, 00:15:35.258 "data_offset": 256, 00:15:35.258 "data_size": 7936 00:15:35.258 }, 00:15:35.258 { 00:15:35.258 "name": "BaseBdev2", 00:15:35.258 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:35.258 "is_configured": true, 00:15:35.258 "data_offset": 256, 00:15:35.258 "data_size": 7936 00:15:35.258 } 00:15:35.258 ] 00:15:35.258 }' 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.258 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:35.258 "name": "raid_bdev1", 00:15:35.258 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:35.258 "strip_size_kb": 0, 00:15:35.258 "state": "online", 00:15:35.258 "raid_level": "raid1", 00:15:35.258 "superblock": true, 00:15:35.258 "num_base_bdevs": 2, 00:15:35.258 "num_base_bdevs_discovered": 2, 00:15:35.258 "num_base_bdevs_operational": 2, 00:15:35.258 "base_bdevs_list": [ 00:15:35.258 { 00:15:35.258 "name": "spare", 00:15:35.258 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:35.258 "is_configured": true, 00:15:35.258 "data_offset": 256, 00:15:35.258 "data_size": 7936 00:15:35.258 }, 00:15:35.258 { 00:15:35.259 "name": "BaseBdev2", 00:15:35.259 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:35.259 "is_configured": true, 00:15:35.259 "data_offset": 256, 00:15:35.259 "data_size": 7936 00:15:35.259 } 00:15:35.259 ] 00:15:35.259 }' 00:15:35.259 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:35.259 16:53:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.828 [2024-09-29 16:53:57.295814] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:35.828 [2024-09-29 16:53:57.295841] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:35.828 [2024-09-29 16:53:57.295916] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:35.828 [2024-09-29 16:53:57.295974] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:35.828 [2024-09-29 16:53:57.295993] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:35.828 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:36.088 /dev/nbd0 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:36.088 1+0 records in 00:15:36.088 1+0 records out 00:15:36.088 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000458804 s, 8.9 MB/s 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:36.088 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:36.348 /dev/nbd1 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:36.348 1+0 records in 00:15:36.348 1+0 records out 00:15:36.348 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000350503 s, 11.7 MB/s 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:36.348 16:53:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:36.607 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:36.866 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:36.866 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:36.866 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:36.866 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:36.866 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.867 [2024-09-29 16:53:58.411573] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:36.867 [2024-09-29 16:53:58.411629] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:36.867 [2024-09-29 16:53:58.411648] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:36.867 [2024-09-29 16:53:58.411660] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:36.867 [2024-09-29 16:53:58.413865] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:36.867 [2024-09-29 16:53:58.413908] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:36.867 [2024-09-29 16:53:58.413977] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:36.867 [2024-09-29 16:53:58.414016] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:36.867 [2024-09-29 16:53:58.414147] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:36.867 spare 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:36.867 [2024-09-29 16:53:58.514059] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:36.867 [2024-09-29 16:53:58.514085] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:36.867 [2024-09-29 16:53:58.514333] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:15:36.867 [2024-09-29 16:53:58.514514] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:36.867 [2024-09-29 16:53:58.514548] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:36.867 [2024-09-29 16:53:58.514684] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.867 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.126 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.126 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.126 "name": "raid_bdev1", 00:15:37.126 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:37.126 "strip_size_kb": 0, 00:15:37.126 "state": "online", 00:15:37.126 "raid_level": "raid1", 00:15:37.126 "superblock": true, 00:15:37.126 "num_base_bdevs": 2, 00:15:37.126 "num_base_bdevs_discovered": 2, 00:15:37.126 "num_base_bdevs_operational": 2, 00:15:37.126 "base_bdevs_list": [ 00:15:37.126 { 00:15:37.126 "name": "spare", 00:15:37.126 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:37.126 "is_configured": true, 00:15:37.126 "data_offset": 256, 00:15:37.126 "data_size": 7936 00:15:37.126 }, 00:15:37.126 { 00:15:37.126 "name": "BaseBdev2", 00:15:37.126 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:37.126 "is_configured": true, 00:15:37.126 "data_offset": 256, 00:15:37.126 "data_size": 7936 00:15:37.126 } 00:15:37.126 ] 00:15:37.126 }' 00:15:37.126 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.126 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:37.386 16:53:58 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.386 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:37.386 "name": "raid_bdev1", 00:15:37.386 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:37.386 "strip_size_kb": 0, 00:15:37.386 "state": "online", 00:15:37.386 "raid_level": "raid1", 00:15:37.386 "superblock": true, 00:15:37.386 "num_base_bdevs": 2, 00:15:37.386 "num_base_bdevs_discovered": 2, 00:15:37.386 "num_base_bdevs_operational": 2, 00:15:37.386 "base_bdevs_list": [ 00:15:37.386 { 00:15:37.386 "name": "spare", 00:15:37.386 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:37.386 "is_configured": true, 00:15:37.386 "data_offset": 256, 00:15:37.386 "data_size": 7936 00:15:37.386 }, 00:15:37.386 { 00:15:37.386 "name": "BaseBdev2", 00:15:37.386 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:37.386 "is_configured": true, 00:15:37.386 "data_offset": 256, 00:15:37.386 "data_size": 7936 00:15:37.386 } 00:15:37.386 ] 00:15:37.386 }' 00:15:37.386 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:37.386 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:37.386 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:37.644 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.645 [2024-09-29 16:53:59.118644] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:37.645 "name": "raid_bdev1", 00:15:37.645 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:37.645 "strip_size_kb": 0, 00:15:37.645 "state": "online", 00:15:37.645 "raid_level": "raid1", 00:15:37.645 "superblock": true, 00:15:37.645 "num_base_bdevs": 2, 00:15:37.645 "num_base_bdevs_discovered": 1, 00:15:37.645 "num_base_bdevs_operational": 1, 00:15:37.645 "base_bdevs_list": [ 00:15:37.645 { 00:15:37.645 "name": null, 00:15:37.645 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:37.645 "is_configured": false, 00:15:37.645 "data_offset": 0, 00:15:37.645 "data_size": 7936 00:15:37.645 }, 00:15:37.645 { 00:15:37.645 "name": "BaseBdev2", 00:15:37.645 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:37.645 "is_configured": true, 00:15:37.645 "data_offset": 256, 00:15:37.645 "data_size": 7936 00:15:37.645 } 00:15:37.645 ] 00:15:37.645 }' 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:37.645 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.215 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:38.215 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.215 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:38.215 [2024-09-29 16:53:59.609835] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:38.215 [2024-09-29 16:53:59.610001] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:38.215 [2024-09-29 16:53:59.610020] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:38.215 [2024-09-29 16:53:59.610052] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:38.215 [2024-09-29 16:53:59.614127] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:15:38.215 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.215 16:53:59 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:38.215 [2024-09-29 16:53:59.616141] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:39.152 "name": "raid_bdev1", 00:15:39.152 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:39.152 "strip_size_kb": 0, 00:15:39.152 "state": "online", 00:15:39.152 "raid_level": "raid1", 00:15:39.152 "superblock": true, 00:15:39.152 "num_base_bdevs": 2, 00:15:39.152 "num_base_bdevs_discovered": 2, 00:15:39.152 "num_base_bdevs_operational": 2, 00:15:39.152 "process": { 00:15:39.152 "type": "rebuild", 00:15:39.152 "target": "spare", 00:15:39.152 "progress": { 00:15:39.152 "blocks": 2560, 00:15:39.152 "percent": 32 00:15:39.152 } 00:15:39.152 }, 00:15:39.152 "base_bdevs_list": [ 00:15:39.152 { 00:15:39.152 "name": "spare", 00:15:39.152 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:39.152 "is_configured": true, 00:15:39.152 "data_offset": 256, 00:15:39.152 "data_size": 7936 00:15:39.152 }, 00:15:39.152 { 00:15:39.152 "name": "BaseBdev2", 00:15:39.152 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:39.152 "is_configured": true, 00:15:39.152 "data_offset": 256, 00:15:39.152 "data_size": 7936 00:15:39.152 } 00:15:39.152 ] 00:15:39.152 }' 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.152 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.152 [2024-09-29 16:54:00.753012] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:39.152 [2024-09-29 16:54:00.820046] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:39.152 [2024-09-29 16:54:00.820116] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:39.152 [2024-09-29 16:54:00.820133] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:39.152 [2024-09-29 16:54:00.820140] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.411 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:39.411 "name": "raid_bdev1", 00:15:39.411 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:39.411 "strip_size_kb": 0, 00:15:39.411 "state": "online", 00:15:39.411 "raid_level": "raid1", 00:15:39.411 "superblock": true, 00:15:39.411 "num_base_bdevs": 2, 00:15:39.411 "num_base_bdevs_discovered": 1, 00:15:39.411 "num_base_bdevs_operational": 1, 00:15:39.411 "base_bdevs_list": [ 00:15:39.412 { 00:15:39.412 "name": null, 00:15:39.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:39.412 "is_configured": false, 00:15:39.412 "data_offset": 0, 00:15:39.412 "data_size": 7936 00:15:39.412 }, 00:15:39.412 { 00:15:39.412 "name": "BaseBdev2", 00:15:39.412 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:39.412 "is_configured": true, 00:15:39.412 "data_offset": 256, 00:15:39.412 "data_size": 7936 00:15:39.412 } 00:15:39.412 ] 00:15:39.412 }' 00:15:39.412 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:39.412 16:54:00 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.669 16:54:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:39.669 16:54:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.669 16:54:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:39.669 [2024-09-29 16:54:01.307379] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:39.669 [2024-09-29 16:54:01.307435] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:39.669 [2024-09-29 16:54:01.307458] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:39.669 [2024-09-29 16:54:01.307467] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:39.669 [2024-09-29 16:54:01.307934] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:39.669 [2024-09-29 16:54:01.307962] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:39.669 [2024-09-29 16:54:01.308042] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:39.669 [2024-09-29 16:54:01.308073] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:39.669 [2024-09-29 16:54:01.308087] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:39.669 [2024-09-29 16:54:01.308107] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:39.669 spare 00:15:39.669 [2024-09-29 16:54:01.311411] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:15:39.669 16:54:01 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.669 16:54:01 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:39.669 [2024-09-29 16:54:01.313289] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:41.046 "name": "raid_bdev1", 00:15:41.046 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:41.046 "strip_size_kb": 0, 00:15:41.046 "state": "online", 00:15:41.046 "raid_level": "raid1", 00:15:41.046 "superblock": true, 00:15:41.046 "num_base_bdevs": 2, 00:15:41.046 "num_base_bdevs_discovered": 2, 00:15:41.046 "num_base_bdevs_operational": 2, 00:15:41.046 "process": { 00:15:41.046 "type": "rebuild", 00:15:41.046 "target": "spare", 00:15:41.046 "progress": { 00:15:41.046 "blocks": 2560, 00:15:41.046 "percent": 32 00:15:41.046 } 00:15:41.046 }, 00:15:41.046 "base_bdevs_list": [ 00:15:41.046 { 00:15:41.046 "name": "spare", 00:15:41.046 "uuid": "caca7783-e5b9-53ef-81dc-7b7add5520fd", 00:15:41.046 "is_configured": true, 00:15:41.046 "data_offset": 256, 00:15:41.046 "data_size": 7936 00:15:41.046 }, 00:15:41.046 { 00:15:41.046 "name": "BaseBdev2", 00:15:41.046 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:41.046 "is_configured": true, 00:15:41.046 "data_offset": 256, 00:15:41.046 "data_size": 7936 00:15:41.046 } 00:15:41.046 ] 00:15:41.046 }' 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.046 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.046 [2024-09-29 16:54:02.459434] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:41.046 [2024-09-29 16:54:02.517230] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:41.046 [2024-09-29 16:54:02.517292] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:41.047 [2024-09-29 16:54:02.517306] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:41.047 [2024-09-29 16:54:02.517315] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:41.047 "name": "raid_bdev1", 00:15:41.047 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:41.047 "strip_size_kb": 0, 00:15:41.047 "state": "online", 00:15:41.047 "raid_level": "raid1", 00:15:41.047 "superblock": true, 00:15:41.047 "num_base_bdevs": 2, 00:15:41.047 "num_base_bdevs_discovered": 1, 00:15:41.047 "num_base_bdevs_operational": 1, 00:15:41.047 "base_bdevs_list": [ 00:15:41.047 { 00:15:41.047 "name": null, 00:15:41.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.047 "is_configured": false, 00:15:41.047 "data_offset": 0, 00:15:41.047 "data_size": 7936 00:15:41.047 }, 00:15:41.047 { 00:15:41.047 "name": "BaseBdev2", 00:15:41.047 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:41.047 "is_configured": true, 00:15:41.047 "data_offset": 256, 00:15:41.047 "data_size": 7936 00:15:41.047 } 00:15:41.047 ] 00:15:41.047 }' 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:41.047 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.615 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:41.616 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:41.616 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:41.616 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:41.616 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:41.616 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.616 16:54:02 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:41.616 "name": "raid_bdev1", 00:15:41.616 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:41.616 "strip_size_kb": 0, 00:15:41.616 "state": "online", 00:15:41.616 "raid_level": "raid1", 00:15:41.616 "superblock": true, 00:15:41.616 "num_base_bdevs": 2, 00:15:41.616 "num_base_bdevs_discovered": 1, 00:15:41.616 "num_base_bdevs_operational": 1, 00:15:41.616 "base_bdevs_list": [ 00:15:41.616 { 00:15:41.616 "name": null, 00:15:41.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:41.616 "is_configured": false, 00:15:41.616 "data_offset": 0, 00:15:41.616 "data_size": 7936 00:15:41.616 }, 00:15:41.616 { 00:15:41.616 "name": "BaseBdev2", 00:15:41.616 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:41.616 "is_configured": true, 00:15:41.616 "data_offset": 256, 00:15:41.616 "data_size": 7936 00:15:41.616 } 00:15:41.616 ] 00:15:41.616 }' 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:41.616 [2024-09-29 16:54:03.140182] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:41.616 [2024-09-29 16:54:03.140235] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:41.616 [2024-09-29 16:54:03.140254] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:41.616 [2024-09-29 16:54:03.140264] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:41.616 [2024-09-29 16:54:03.140681] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:41.616 [2024-09-29 16:54:03.140712] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:41.616 [2024-09-29 16:54:03.140798] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:41.616 [2024-09-29 16:54:03.140819] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:41.616 [2024-09-29 16:54:03.140829] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:41.616 [2024-09-29 16:54:03.140839] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:41.616 BaseBdev1 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.616 16:54:03 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:42.554 "name": "raid_bdev1", 00:15:42.554 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:42.554 "strip_size_kb": 0, 00:15:42.554 "state": "online", 00:15:42.554 "raid_level": "raid1", 00:15:42.554 "superblock": true, 00:15:42.554 "num_base_bdevs": 2, 00:15:42.554 "num_base_bdevs_discovered": 1, 00:15:42.554 "num_base_bdevs_operational": 1, 00:15:42.554 "base_bdevs_list": [ 00:15:42.554 { 00:15:42.554 "name": null, 00:15:42.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:42.554 "is_configured": false, 00:15:42.554 "data_offset": 0, 00:15:42.554 "data_size": 7936 00:15:42.554 }, 00:15:42.554 { 00:15:42.554 "name": "BaseBdev2", 00:15:42.554 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:42.554 "is_configured": true, 00:15:42.554 "data_offset": 256, 00:15:42.554 "data_size": 7936 00:15:42.554 } 00:15:42.554 ] 00:15:42.554 }' 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:42.554 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.122 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:43.122 "name": "raid_bdev1", 00:15:43.122 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:43.122 "strip_size_kb": 0, 00:15:43.122 "state": "online", 00:15:43.122 "raid_level": "raid1", 00:15:43.122 "superblock": true, 00:15:43.122 "num_base_bdevs": 2, 00:15:43.122 "num_base_bdevs_discovered": 1, 00:15:43.122 "num_base_bdevs_operational": 1, 00:15:43.122 "base_bdevs_list": [ 00:15:43.122 { 00:15:43.122 "name": null, 00:15:43.122 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:43.122 "is_configured": false, 00:15:43.122 "data_offset": 0, 00:15:43.122 "data_size": 7936 00:15:43.123 }, 00:15:43.123 { 00:15:43.123 "name": "BaseBdev2", 00:15:43.123 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:43.123 "is_configured": true, 00:15:43.123 "data_offset": 256, 00:15:43.123 "data_size": 7936 00:15:43.123 } 00:15:43.123 ] 00:15:43.123 }' 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@650 -- # local es=0 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:43.123 [2024-09-29 16:54:04.781439] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:43.123 [2024-09-29 16:54:04.781574] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:43.123 [2024-09-29 16:54:04.781588] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:43.123 request: 00:15:43.123 { 00:15:43.123 "base_bdev": "BaseBdev1", 00:15:43.123 "raid_bdev": "raid_bdev1", 00:15:43.123 "method": "bdev_raid_add_base_bdev", 00:15:43.123 "req_id": 1 00:15:43.123 } 00:15:43.123 Got JSON-RPC error response 00:15:43.123 response: 00:15:43.123 { 00:15:43.123 "code": -22, 00:15:43.123 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:43.123 } 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # es=1 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:43.123 16:54:04 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:44.500 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:44.500 "name": "raid_bdev1", 00:15:44.500 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:44.500 "strip_size_kb": 0, 00:15:44.500 "state": "online", 00:15:44.500 "raid_level": "raid1", 00:15:44.500 "superblock": true, 00:15:44.500 "num_base_bdevs": 2, 00:15:44.500 "num_base_bdevs_discovered": 1, 00:15:44.500 "num_base_bdevs_operational": 1, 00:15:44.500 "base_bdevs_list": [ 00:15:44.500 { 00:15:44.500 "name": null, 00:15:44.500 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.501 "is_configured": false, 00:15:44.501 "data_offset": 0, 00:15:44.501 "data_size": 7936 00:15:44.501 }, 00:15:44.501 { 00:15:44.501 "name": "BaseBdev2", 00:15:44.501 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:44.501 "is_configured": true, 00:15:44.501 "data_offset": 256, 00:15:44.501 "data_size": 7936 00:15:44.501 } 00:15:44.501 ] 00:15:44.501 }' 00:15:44.501 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:44.501 16:54:05 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:44.769 "name": "raid_bdev1", 00:15:44.769 "uuid": "0cfec959-40e9-46b1-ab8e-5e771c04b1bc", 00:15:44.769 "strip_size_kb": 0, 00:15:44.769 "state": "online", 00:15:44.769 "raid_level": "raid1", 00:15:44.769 "superblock": true, 00:15:44.769 "num_base_bdevs": 2, 00:15:44.769 "num_base_bdevs_discovered": 1, 00:15:44.769 "num_base_bdevs_operational": 1, 00:15:44.769 "base_bdevs_list": [ 00:15:44.769 { 00:15:44.769 "name": null, 00:15:44.769 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:44.769 "is_configured": false, 00:15:44.769 "data_offset": 0, 00:15:44.769 "data_size": 7936 00:15:44.769 }, 00:15:44.769 { 00:15:44.769 "name": "BaseBdev2", 00:15:44.769 "uuid": "ab900271-d0a7-5464-95c8-5a16b25659ca", 00:15:44.769 "is_configured": true, 00:15:44.769 "data_offset": 256, 00:15:44.769 "data_size": 7936 00:15:44.769 } 00:15:44.769 ] 00:15:44.769 }' 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 96503 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 96503 ']' 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 96503 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:44.769 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 96503 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:45.051 killing process with pid 96503 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 96503' 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@969 -- # kill 96503 00:15:45.051 Received shutdown signal, test time was about 60.000000 seconds 00:15:45.051 00:15:45.051 Latency(us) 00:15:45.051 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:45.051 =================================================================================================================== 00:15:45.051 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:45.051 [2024-09-29 16:54:06.453780] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:45.051 [2024-09-29 16:54:06.453924] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@974 -- # wait 96503 00:15:45.051 [2024-09-29 16:54:06.453977] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:45.051 [2024-09-29 16:54:06.453986] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:45.051 [2024-09-29 16:54:06.485667] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:15:45.051 00:15:45.051 real 0m18.563s 00:15:45.051 user 0m24.676s 00:15:45.051 sys 0m2.701s 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:45.051 16:54:06 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:45.051 ************************************ 00:15:45.051 END TEST raid_rebuild_test_sb_4k 00:15:45.051 ************************************ 00:15:45.330 16:54:06 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:15:45.330 16:54:06 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:15:45.330 16:54:06 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:15:45.330 16:54:06 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:45.330 16:54:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:45.330 ************************************ 00:15:45.330 START TEST raid_state_function_test_sb_md_separate 00:15:45.330 ************************************ 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:45.330 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=97185 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:45.331 Process raid pid: 97185 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97185' 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 97185 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 97185 ']' 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:45.331 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:45.331 16:54:06 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:45.331 [2024-09-29 16:54:06.897022] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:15:45.331 [2024-09-29 16:54:06.897211] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:45.591 [2024-09-29 16:54:07.043649] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:45.591 [2024-09-29 16:54:07.089462] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:45.591 [2024-09-29 16:54:07.131944] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:45.591 [2024-09-29 16:54:07.131980] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.161 [2024-09-29 16:54:07.737394] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:46.161 [2024-09-29 16:54:07.737447] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:46.161 [2024-09-29 16:54:07.737459] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:46.161 [2024-09-29 16:54:07.737469] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.161 "name": "Existed_Raid", 00:15:46.161 "uuid": "ff88cc14-8a06-40c6-93b7-f131480c8021", 00:15:46.161 "strip_size_kb": 0, 00:15:46.161 "state": "configuring", 00:15:46.161 "raid_level": "raid1", 00:15:46.161 "superblock": true, 00:15:46.161 "num_base_bdevs": 2, 00:15:46.161 "num_base_bdevs_discovered": 0, 00:15:46.161 "num_base_bdevs_operational": 2, 00:15:46.161 "base_bdevs_list": [ 00:15:46.161 { 00:15:46.161 "name": "BaseBdev1", 00:15:46.161 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.161 "is_configured": false, 00:15:46.161 "data_offset": 0, 00:15:46.161 "data_size": 0 00:15:46.161 }, 00:15:46.161 { 00:15:46.161 "name": "BaseBdev2", 00:15:46.161 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.161 "is_configured": false, 00:15:46.161 "data_offset": 0, 00:15:46.161 "data_size": 0 00:15:46.161 } 00:15:46.161 ] 00:15:46.161 }' 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.161 16:54:07 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.730 [2024-09-29 16:54:08.196552] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:46.730 [2024-09-29 16:54:08.196593] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.730 [2024-09-29 16:54:08.208561] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:46.730 [2024-09-29 16:54:08.208598] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:46.730 [2024-09-29 16:54:08.208613] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:46.730 [2024-09-29 16:54:08.208623] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.730 [2024-09-29 16:54:08.230056] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:46.730 BaseBdev1 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.730 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.730 [ 00:15:46.730 { 00:15:46.730 "name": "BaseBdev1", 00:15:46.730 "aliases": [ 00:15:46.730 "afbc9c11-3045-46c6-8a10-9a454d7fbf71" 00:15:46.730 ], 00:15:46.730 "product_name": "Malloc disk", 00:15:46.730 "block_size": 4096, 00:15:46.730 "num_blocks": 8192, 00:15:46.730 "uuid": "afbc9c11-3045-46c6-8a10-9a454d7fbf71", 00:15:46.730 "md_size": 32, 00:15:46.730 "md_interleave": false, 00:15:46.730 "dif_type": 0, 00:15:46.730 "assigned_rate_limits": { 00:15:46.730 "rw_ios_per_sec": 0, 00:15:46.730 "rw_mbytes_per_sec": 0, 00:15:46.730 "r_mbytes_per_sec": 0, 00:15:46.730 "w_mbytes_per_sec": 0 00:15:46.730 }, 00:15:46.730 "claimed": true, 00:15:46.730 "claim_type": "exclusive_write", 00:15:46.730 "zoned": false, 00:15:46.730 "supported_io_types": { 00:15:46.730 "read": true, 00:15:46.730 "write": true, 00:15:46.730 "unmap": true, 00:15:46.730 "flush": true, 00:15:46.730 "reset": true, 00:15:46.730 "nvme_admin": false, 00:15:46.730 "nvme_io": false, 00:15:46.730 "nvme_io_md": false, 00:15:46.730 "write_zeroes": true, 00:15:46.730 "zcopy": true, 00:15:46.730 "get_zone_info": false, 00:15:46.730 "zone_management": false, 00:15:46.730 "zone_append": false, 00:15:46.730 "compare": false, 00:15:46.730 "compare_and_write": false, 00:15:46.730 "abort": true, 00:15:46.730 "seek_hole": false, 00:15:46.730 "seek_data": false, 00:15:46.730 "copy": true, 00:15:46.730 "nvme_iov_md": false 00:15:46.730 }, 00:15:46.730 "memory_domains": [ 00:15:46.730 { 00:15:46.730 "dma_device_id": "system", 00:15:46.730 "dma_device_type": 1 00:15:46.730 }, 00:15:46.730 { 00:15:46.730 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:46.730 "dma_device_type": 2 00:15:46.730 } 00:15:46.730 ], 00:15:46.730 "driver_specific": {} 00:15:46.731 } 00:15:46.731 ] 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:46.731 "name": "Existed_Raid", 00:15:46.731 "uuid": "30a62ff4-bf71-4af0-8cfd-71ed59d1afc2", 00:15:46.731 "strip_size_kb": 0, 00:15:46.731 "state": "configuring", 00:15:46.731 "raid_level": "raid1", 00:15:46.731 "superblock": true, 00:15:46.731 "num_base_bdevs": 2, 00:15:46.731 "num_base_bdevs_discovered": 1, 00:15:46.731 "num_base_bdevs_operational": 2, 00:15:46.731 "base_bdevs_list": [ 00:15:46.731 { 00:15:46.731 "name": "BaseBdev1", 00:15:46.731 "uuid": "afbc9c11-3045-46c6-8a10-9a454d7fbf71", 00:15:46.731 "is_configured": true, 00:15:46.731 "data_offset": 256, 00:15:46.731 "data_size": 7936 00:15:46.731 }, 00:15:46.731 { 00:15:46.731 "name": "BaseBdev2", 00:15:46.731 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:46.731 "is_configured": false, 00:15:46.731 "data_offset": 0, 00:15:46.731 "data_size": 0 00:15:46.731 } 00:15:46.731 ] 00:15:46.731 }' 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:46.731 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.300 [2024-09-29 16:54:08.765197] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:47.300 [2024-09-29 16:54:08.765239] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.300 [2024-09-29 16:54:08.777224] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:47.300 [2024-09-29 16:54:08.779049] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:47.300 [2024-09-29 16:54:08.779086] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:47.300 "name": "Existed_Raid", 00:15:47.300 "uuid": "e9176b32-6246-4161-998f-0c22d5885bd7", 00:15:47.300 "strip_size_kb": 0, 00:15:47.300 "state": "configuring", 00:15:47.300 "raid_level": "raid1", 00:15:47.300 "superblock": true, 00:15:47.300 "num_base_bdevs": 2, 00:15:47.300 "num_base_bdevs_discovered": 1, 00:15:47.300 "num_base_bdevs_operational": 2, 00:15:47.300 "base_bdevs_list": [ 00:15:47.300 { 00:15:47.300 "name": "BaseBdev1", 00:15:47.300 "uuid": "afbc9c11-3045-46c6-8a10-9a454d7fbf71", 00:15:47.300 "is_configured": true, 00:15:47.300 "data_offset": 256, 00:15:47.300 "data_size": 7936 00:15:47.300 }, 00:15:47.300 { 00:15:47.300 "name": "BaseBdev2", 00:15:47.300 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:47.300 "is_configured": false, 00:15:47.300 "data_offset": 0, 00:15:47.300 "data_size": 0 00:15:47.300 } 00:15:47.300 ] 00:15:47.300 }' 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:47.300 16:54:08 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.560 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:15:47.560 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.560 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.820 [2024-09-29 16:54:09.255362] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:47.820 [2024-09-29 16:54:09.256031] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:47.820 [2024-09-29 16:54:09.256095] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:47.820 [2024-09-29 16:54:09.256431] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:47.820 BaseBdev2 00:15:47.820 [2024-09-29 16:54:09.256953] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:47.820 [2024-09-29 16:54:09.257032] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.820 [2024-09-29 16:54:09.257368] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.820 [ 00:15:47.820 { 00:15:47.820 "name": "BaseBdev2", 00:15:47.820 "aliases": [ 00:15:47.820 "79223097-9afe-4616-a6b5-f1dc47e6c975" 00:15:47.820 ], 00:15:47.820 "product_name": "Malloc disk", 00:15:47.820 "block_size": 4096, 00:15:47.820 "num_blocks": 8192, 00:15:47.820 "uuid": "79223097-9afe-4616-a6b5-f1dc47e6c975", 00:15:47.820 "md_size": 32, 00:15:47.820 "md_interleave": false, 00:15:47.820 "dif_type": 0, 00:15:47.820 "assigned_rate_limits": { 00:15:47.820 "rw_ios_per_sec": 0, 00:15:47.820 "rw_mbytes_per_sec": 0, 00:15:47.820 "r_mbytes_per_sec": 0, 00:15:47.820 "w_mbytes_per_sec": 0 00:15:47.820 }, 00:15:47.820 "claimed": true, 00:15:47.820 "claim_type": "exclusive_write", 00:15:47.820 "zoned": false, 00:15:47.820 "supported_io_types": { 00:15:47.820 "read": true, 00:15:47.820 "write": true, 00:15:47.820 "unmap": true, 00:15:47.820 "flush": true, 00:15:47.820 "reset": true, 00:15:47.820 "nvme_admin": false, 00:15:47.820 "nvme_io": false, 00:15:47.820 "nvme_io_md": false, 00:15:47.820 "write_zeroes": true, 00:15:47.820 "zcopy": true, 00:15:47.820 "get_zone_info": false, 00:15:47.820 "zone_management": false, 00:15:47.820 "zone_append": false, 00:15:47.820 "compare": false, 00:15:47.820 "compare_and_write": false, 00:15:47.820 "abort": true, 00:15:47.820 "seek_hole": false, 00:15:47.820 "seek_data": false, 00:15:47.820 "copy": true, 00:15:47.820 "nvme_iov_md": false 00:15:47.820 }, 00:15:47.820 "memory_domains": [ 00:15:47.820 { 00:15:47.820 "dma_device_id": "system", 00:15:47.820 "dma_device_type": 1 00:15:47.820 }, 00:15:47.820 { 00:15:47.820 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:47.820 "dma_device_type": 2 00:15:47.820 } 00:15:47.820 ], 00:15:47.820 "driver_specific": {} 00:15:47.820 } 00:15:47.820 ] 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:47.820 "name": "Existed_Raid", 00:15:47.820 "uuid": "e9176b32-6246-4161-998f-0c22d5885bd7", 00:15:47.820 "strip_size_kb": 0, 00:15:47.820 "state": "online", 00:15:47.820 "raid_level": "raid1", 00:15:47.820 "superblock": true, 00:15:47.820 "num_base_bdevs": 2, 00:15:47.820 "num_base_bdevs_discovered": 2, 00:15:47.820 "num_base_bdevs_operational": 2, 00:15:47.820 "base_bdevs_list": [ 00:15:47.820 { 00:15:47.820 "name": "BaseBdev1", 00:15:47.820 "uuid": "afbc9c11-3045-46c6-8a10-9a454d7fbf71", 00:15:47.820 "is_configured": true, 00:15:47.820 "data_offset": 256, 00:15:47.820 "data_size": 7936 00:15:47.820 }, 00:15:47.820 { 00:15:47.820 "name": "BaseBdev2", 00:15:47.820 "uuid": "79223097-9afe-4616-a6b5-f1dc47e6c975", 00:15:47.820 "is_configured": true, 00:15:47.820 "data_offset": 256, 00:15:47.820 "data_size": 7936 00:15:47.820 } 00:15:47.820 ] 00:15:47.820 }' 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:47.820 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.391 [2024-09-29 16:54:09.774714] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:48.391 "name": "Existed_Raid", 00:15:48.391 "aliases": [ 00:15:48.391 "e9176b32-6246-4161-998f-0c22d5885bd7" 00:15:48.391 ], 00:15:48.391 "product_name": "Raid Volume", 00:15:48.391 "block_size": 4096, 00:15:48.391 "num_blocks": 7936, 00:15:48.391 "uuid": "e9176b32-6246-4161-998f-0c22d5885bd7", 00:15:48.391 "md_size": 32, 00:15:48.391 "md_interleave": false, 00:15:48.391 "dif_type": 0, 00:15:48.391 "assigned_rate_limits": { 00:15:48.391 "rw_ios_per_sec": 0, 00:15:48.391 "rw_mbytes_per_sec": 0, 00:15:48.391 "r_mbytes_per_sec": 0, 00:15:48.391 "w_mbytes_per_sec": 0 00:15:48.391 }, 00:15:48.391 "claimed": false, 00:15:48.391 "zoned": false, 00:15:48.391 "supported_io_types": { 00:15:48.391 "read": true, 00:15:48.391 "write": true, 00:15:48.391 "unmap": false, 00:15:48.391 "flush": false, 00:15:48.391 "reset": true, 00:15:48.391 "nvme_admin": false, 00:15:48.391 "nvme_io": false, 00:15:48.391 "nvme_io_md": false, 00:15:48.391 "write_zeroes": true, 00:15:48.391 "zcopy": false, 00:15:48.391 "get_zone_info": false, 00:15:48.391 "zone_management": false, 00:15:48.391 "zone_append": false, 00:15:48.391 "compare": false, 00:15:48.391 "compare_and_write": false, 00:15:48.391 "abort": false, 00:15:48.391 "seek_hole": false, 00:15:48.391 "seek_data": false, 00:15:48.391 "copy": false, 00:15:48.391 "nvme_iov_md": false 00:15:48.391 }, 00:15:48.391 "memory_domains": [ 00:15:48.391 { 00:15:48.391 "dma_device_id": "system", 00:15:48.391 "dma_device_type": 1 00:15:48.391 }, 00:15:48.391 { 00:15:48.391 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:48.391 "dma_device_type": 2 00:15:48.391 }, 00:15:48.391 { 00:15:48.391 "dma_device_id": "system", 00:15:48.391 "dma_device_type": 1 00:15:48.391 }, 00:15:48.391 { 00:15:48.391 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:48.391 "dma_device_type": 2 00:15:48.391 } 00:15:48.391 ], 00:15:48.391 "driver_specific": { 00:15:48.391 "raid": { 00:15:48.391 "uuid": "e9176b32-6246-4161-998f-0c22d5885bd7", 00:15:48.391 "strip_size_kb": 0, 00:15:48.391 "state": "online", 00:15:48.391 "raid_level": "raid1", 00:15:48.391 "superblock": true, 00:15:48.391 "num_base_bdevs": 2, 00:15:48.391 "num_base_bdevs_discovered": 2, 00:15:48.391 "num_base_bdevs_operational": 2, 00:15:48.391 "base_bdevs_list": [ 00:15:48.391 { 00:15:48.391 "name": "BaseBdev1", 00:15:48.391 "uuid": "afbc9c11-3045-46c6-8a10-9a454d7fbf71", 00:15:48.391 "is_configured": true, 00:15:48.391 "data_offset": 256, 00:15:48.391 "data_size": 7936 00:15:48.391 }, 00:15:48.391 { 00:15:48.391 "name": "BaseBdev2", 00:15:48.391 "uuid": "79223097-9afe-4616-a6b5-f1dc47e6c975", 00:15:48.391 "is_configured": true, 00:15:48.391 "data_offset": 256, 00:15:48.391 "data_size": 7936 00:15:48.391 } 00:15:48.391 ] 00:15:48.391 } 00:15:48.391 } 00:15:48.391 }' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:48.391 BaseBdev2' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.391 [2024-09-29 16:54:09.966191] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:48.391 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:48.392 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:48.392 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.392 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.392 16:54:09 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.392 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.392 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:48.392 "name": "Existed_Raid", 00:15:48.392 "uuid": "e9176b32-6246-4161-998f-0c22d5885bd7", 00:15:48.392 "strip_size_kb": 0, 00:15:48.392 "state": "online", 00:15:48.392 "raid_level": "raid1", 00:15:48.392 "superblock": true, 00:15:48.392 "num_base_bdevs": 2, 00:15:48.392 "num_base_bdevs_discovered": 1, 00:15:48.392 "num_base_bdevs_operational": 1, 00:15:48.392 "base_bdevs_list": [ 00:15:48.392 { 00:15:48.392 "name": null, 00:15:48.392 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:48.392 "is_configured": false, 00:15:48.392 "data_offset": 0, 00:15:48.392 "data_size": 7936 00:15:48.392 }, 00:15:48.392 { 00:15:48.392 "name": "BaseBdev2", 00:15:48.392 "uuid": "79223097-9afe-4616-a6b5-f1dc47e6c975", 00:15:48.392 "is_configured": true, 00:15:48.392 "data_offset": 256, 00:15:48.392 "data_size": 7936 00:15:48.392 } 00:15:48.392 ] 00:15:48.392 }' 00:15:48.392 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:48.392 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.962 [2024-09-29 16:54:10.461495] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:48.962 [2024-09-29 16:54:10.461597] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:48.962 [2024-09-29 16:54:10.473785] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:48.962 [2024-09-29 16:54:10.473827] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:48.962 [2024-09-29 16:54:10.473838] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 97185 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 97185 ']' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 97185 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97185 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:48.962 killing process with pid 97185 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97185' 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 97185 00:15:48.962 [2024-09-29 16:54:10.561151] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:48.962 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 97185 00:15:48.962 [2024-09-29 16:54:10.562122] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:49.223 ************************************ 00:15:49.223 END TEST raid_state_function_test_sb_md_separate 00:15:49.223 ************************************ 00:15:49.223 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:15:49.223 00:15:49.223 real 0m4.014s 00:15:49.223 user 0m6.239s 00:15:49.223 sys 0m0.924s 00:15:49.223 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:49.223 16:54:10 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.223 16:54:10 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:15:49.223 16:54:10 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:15:49.223 16:54:10 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:49.223 16:54:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:49.223 ************************************ 00:15:49.223 START TEST raid_superblock_test_md_separate 00:15:49.223 ************************************ 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=97425 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 97425 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@831 -- # '[' -z 97425 ']' 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:49.223 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:49.223 16:54:10 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:49.483 [2024-09-29 16:54:10.977191] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:15:49.483 [2024-09-29 16:54:10.977313] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97425 ] 00:15:49.483 [2024-09-29 16:54:11.123030] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:49.743 [2024-09-29 16:54:11.171658] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:49.743 [2024-09-29 16:54:11.214553] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:49.743 [2024-09-29 16:54:11.214616] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # return 0 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:15:50.314 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.315 malloc1 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.315 [2024-09-29 16:54:11.813820] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:50.315 [2024-09-29 16:54:11.813883] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:50.315 [2024-09-29 16:54:11.813903] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:50.315 [2024-09-29 16:54:11.813914] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:50.315 [2024-09-29 16:54:11.815942] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:50.315 [2024-09-29 16:54:11.815979] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:50.315 pt1 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.315 malloc2 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.315 [2024-09-29 16:54:11.856201] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:50.315 [2024-09-29 16:54:11.856253] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:50.315 [2024-09-29 16:54:11.856270] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:50.315 [2024-09-29 16:54:11.856280] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:50.315 [2024-09-29 16:54:11.858168] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:50.315 [2024-09-29 16:54:11.858202] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:50.315 pt2 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.315 [2024-09-29 16:54:11.868225] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:50.315 [2024-09-29 16:54:11.870170] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:50.315 [2024-09-29 16:54:11.870326] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:50.315 [2024-09-29 16:54:11.870342] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:50.315 [2024-09-29 16:54:11.870420] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:15:50.315 [2024-09-29 16:54:11.870534] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:50.315 [2024-09-29 16:54:11.870548] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:50.315 [2024-09-29 16:54:11.870625] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:50.315 "name": "raid_bdev1", 00:15:50.315 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:50.315 "strip_size_kb": 0, 00:15:50.315 "state": "online", 00:15:50.315 "raid_level": "raid1", 00:15:50.315 "superblock": true, 00:15:50.315 "num_base_bdevs": 2, 00:15:50.315 "num_base_bdevs_discovered": 2, 00:15:50.315 "num_base_bdevs_operational": 2, 00:15:50.315 "base_bdevs_list": [ 00:15:50.315 { 00:15:50.315 "name": "pt1", 00:15:50.315 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:50.315 "is_configured": true, 00:15:50.315 "data_offset": 256, 00:15:50.315 "data_size": 7936 00:15:50.315 }, 00:15:50.315 { 00:15:50.315 "name": "pt2", 00:15:50.315 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:50.315 "is_configured": true, 00:15:50.315 "data_offset": 256, 00:15:50.315 "data_size": 7936 00:15:50.315 } 00:15:50.315 ] 00:15:50.315 }' 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:50.315 16:54:11 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:50.886 [2024-09-29 16:54:12.347697] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:50.886 "name": "raid_bdev1", 00:15:50.886 "aliases": [ 00:15:50.886 "41d985d4-7f8d-4d10-8f0b-1fe972d845a7" 00:15:50.886 ], 00:15:50.886 "product_name": "Raid Volume", 00:15:50.886 "block_size": 4096, 00:15:50.886 "num_blocks": 7936, 00:15:50.886 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:50.886 "md_size": 32, 00:15:50.886 "md_interleave": false, 00:15:50.886 "dif_type": 0, 00:15:50.886 "assigned_rate_limits": { 00:15:50.886 "rw_ios_per_sec": 0, 00:15:50.886 "rw_mbytes_per_sec": 0, 00:15:50.886 "r_mbytes_per_sec": 0, 00:15:50.886 "w_mbytes_per_sec": 0 00:15:50.886 }, 00:15:50.886 "claimed": false, 00:15:50.886 "zoned": false, 00:15:50.886 "supported_io_types": { 00:15:50.886 "read": true, 00:15:50.886 "write": true, 00:15:50.886 "unmap": false, 00:15:50.886 "flush": false, 00:15:50.886 "reset": true, 00:15:50.886 "nvme_admin": false, 00:15:50.886 "nvme_io": false, 00:15:50.886 "nvme_io_md": false, 00:15:50.886 "write_zeroes": true, 00:15:50.886 "zcopy": false, 00:15:50.886 "get_zone_info": false, 00:15:50.886 "zone_management": false, 00:15:50.886 "zone_append": false, 00:15:50.886 "compare": false, 00:15:50.886 "compare_and_write": false, 00:15:50.886 "abort": false, 00:15:50.886 "seek_hole": false, 00:15:50.886 "seek_data": false, 00:15:50.886 "copy": false, 00:15:50.886 "nvme_iov_md": false 00:15:50.886 }, 00:15:50.886 "memory_domains": [ 00:15:50.886 { 00:15:50.886 "dma_device_id": "system", 00:15:50.886 "dma_device_type": 1 00:15:50.886 }, 00:15:50.886 { 00:15:50.886 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.886 "dma_device_type": 2 00:15:50.886 }, 00:15:50.886 { 00:15:50.886 "dma_device_id": "system", 00:15:50.886 "dma_device_type": 1 00:15:50.886 }, 00:15:50.886 { 00:15:50.886 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:50.886 "dma_device_type": 2 00:15:50.886 } 00:15:50.886 ], 00:15:50.886 "driver_specific": { 00:15:50.886 "raid": { 00:15:50.886 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:50.886 "strip_size_kb": 0, 00:15:50.886 "state": "online", 00:15:50.886 "raid_level": "raid1", 00:15:50.886 "superblock": true, 00:15:50.886 "num_base_bdevs": 2, 00:15:50.886 "num_base_bdevs_discovered": 2, 00:15:50.886 "num_base_bdevs_operational": 2, 00:15:50.886 "base_bdevs_list": [ 00:15:50.886 { 00:15:50.886 "name": "pt1", 00:15:50.886 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:50.886 "is_configured": true, 00:15:50.886 "data_offset": 256, 00:15:50.886 "data_size": 7936 00:15:50.886 }, 00:15:50.886 { 00:15:50.886 "name": "pt2", 00:15:50.886 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:50.886 "is_configured": true, 00:15:50.886 "data_offset": 256, 00:15:50.886 "data_size": 7936 00:15:50.886 } 00:15:50.886 ] 00:15:50.886 } 00:15:50.886 } 00:15:50.886 }' 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:50.886 pt2' 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:50.886 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.147 [2024-09-29 16:54:12.583247] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=41d985d4-7f8d-4d10-8f0b-1fe972d845a7 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z 41d985d4-7f8d-4d10-8f0b-1fe972d845a7 ']' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.147 [2024-09-29 16:54:12.610993] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:51.147 [2024-09-29 16:54:12.611018] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:51.147 [2024-09-29 16:54:12.611085] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:51.147 [2024-09-29 16:54:12.611136] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:51.147 [2024-09-29 16:54:12.611146] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:51.147 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.148 [2024-09-29 16:54:12.754826] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:51.148 [2024-09-29 16:54:12.756689] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:51.148 [2024-09-29 16:54:12.756759] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:51.148 [2024-09-29 16:54:12.756799] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:51.148 [2024-09-29 16:54:12.756816] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:51.148 [2024-09-29 16:54:12.756825] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:51.148 request: 00:15:51.148 { 00:15:51.148 "name": "raid_bdev1", 00:15:51.148 "raid_level": "raid1", 00:15:51.148 "base_bdevs": [ 00:15:51.148 "malloc1", 00:15:51.148 "malloc2" 00:15:51.148 ], 00:15:51.148 "superblock": false, 00:15:51.148 "method": "bdev_raid_create", 00:15:51.148 "req_id": 1 00:15:51.148 } 00:15:51.148 Got JSON-RPC error response 00:15:51.148 response: 00:15:51.148 { 00:15:51.148 "code": -17, 00:15:51.148 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:51.148 } 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # es=1 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.148 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.148 [2024-09-29 16:54:12.818690] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:51.148 [2024-09-29 16:54:12.818752] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:51.148 [2024-09-29 16:54:12.818772] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:51.148 [2024-09-29 16:54:12.818780] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:51.408 [2024-09-29 16:54:12.820620] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:51.408 [2024-09-29 16:54:12.820661] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:51.408 [2024-09-29 16:54:12.820701] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:51.408 [2024-09-29 16:54:12.820747] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:51.408 pt1 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.408 "name": "raid_bdev1", 00:15:51.408 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:51.408 "strip_size_kb": 0, 00:15:51.408 "state": "configuring", 00:15:51.408 "raid_level": "raid1", 00:15:51.408 "superblock": true, 00:15:51.408 "num_base_bdevs": 2, 00:15:51.408 "num_base_bdevs_discovered": 1, 00:15:51.408 "num_base_bdevs_operational": 2, 00:15:51.408 "base_bdevs_list": [ 00:15:51.408 { 00:15:51.408 "name": "pt1", 00:15:51.408 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:51.408 "is_configured": true, 00:15:51.408 "data_offset": 256, 00:15:51.408 "data_size": 7936 00:15:51.408 }, 00:15:51.408 { 00:15:51.408 "name": null, 00:15:51.408 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:51.408 "is_configured": false, 00:15:51.408 "data_offset": 256, 00:15:51.408 "data_size": 7936 00:15:51.408 } 00:15:51.408 ] 00:15:51.408 }' 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.408 16:54:12 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.668 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:15:51.668 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:51.668 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:51.668 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:51.668 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.668 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.668 [2024-09-29 16:54:13.269914] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:51.668 [2024-09-29 16:54:13.269954] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:51.668 [2024-09-29 16:54:13.269973] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:51.668 [2024-09-29 16:54:13.269980] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:51.668 [2024-09-29 16:54:13.270115] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:51.668 [2024-09-29 16:54:13.270129] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:51.668 [2024-09-29 16:54:13.270164] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:51.668 [2024-09-29 16:54:13.270186] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:51.668 [2024-09-29 16:54:13.270258] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:51.668 [2024-09-29 16:54:13.270265] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:51.668 [2024-09-29 16:54:13.270332] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:51.668 [2024-09-29 16:54:13.270401] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:51.669 [2024-09-29 16:54:13.270412] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:51.669 [2024-09-29 16:54:13.270463] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:51.669 pt2 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.669 "name": "raid_bdev1", 00:15:51.669 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:51.669 "strip_size_kb": 0, 00:15:51.669 "state": "online", 00:15:51.669 "raid_level": "raid1", 00:15:51.669 "superblock": true, 00:15:51.669 "num_base_bdevs": 2, 00:15:51.669 "num_base_bdevs_discovered": 2, 00:15:51.669 "num_base_bdevs_operational": 2, 00:15:51.669 "base_bdevs_list": [ 00:15:51.669 { 00:15:51.669 "name": "pt1", 00:15:51.669 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:51.669 "is_configured": true, 00:15:51.669 "data_offset": 256, 00:15:51.669 "data_size": 7936 00:15:51.669 }, 00:15:51.669 { 00:15:51.669 "name": "pt2", 00:15:51.669 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:51.669 "is_configured": true, 00:15:51.669 "data_offset": 256, 00:15:51.669 "data_size": 7936 00:15:51.669 } 00:15:51.669 ] 00:15:51.669 }' 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.669 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.239 [2024-09-29 16:54:13.673443] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.239 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:52.239 "name": "raid_bdev1", 00:15:52.239 "aliases": [ 00:15:52.239 "41d985d4-7f8d-4d10-8f0b-1fe972d845a7" 00:15:52.239 ], 00:15:52.239 "product_name": "Raid Volume", 00:15:52.239 "block_size": 4096, 00:15:52.239 "num_blocks": 7936, 00:15:52.239 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:52.239 "md_size": 32, 00:15:52.239 "md_interleave": false, 00:15:52.239 "dif_type": 0, 00:15:52.239 "assigned_rate_limits": { 00:15:52.239 "rw_ios_per_sec": 0, 00:15:52.239 "rw_mbytes_per_sec": 0, 00:15:52.239 "r_mbytes_per_sec": 0, 00:15:52.239 "w_mbytes_per_sec": 0 00:15:52.239 }, 00:15:52.239 "claimed": false, 00:15:52.239 "zoned": false, 00:15:52.239 "supported_io_types": { 00:15:52.239 "read": true, 00:15:52.239 "write": true, 00:15:52.239 "unmap": false, 00:15:52.240 "flush": false, 00:15:52.240 "reset": true, 00:15:52.240 "nvme_admin": false, 00:15:52.240 "nvme_io": false, 00:15:52.240 "nvme_io_md": false, 00:15:52.240 "write_zeroes": true, 00:15:52.240 "zcopy": false, 00:15:52.240 "get_zone_info": false, 00:15:52.240 "zone_management": false, 00:15:52.240 "zone_append": false, 00:15:52.240 "compare": false, 00:15:52.240 "compare_and_write": false, 00:15:52.240 "abort": false, 00:15:52.240 "seek_hole": false, 00:15:52.240 "seek_data": false, 00:15:52.240 "copy": false, 00:15:52.240 "nvme_iov_md": false 00:15:52.240 }, 00:15:52.240 "memory_domains": [ 00:15:52.240 { 00:15:52.240 "dma_device_id": "system", 00:15:52.240 "dma_device_type": 1 00:15:52.240 }, 00:15:52.240 { 00:15:52.240 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:52.240 "dma_device_type": 2 00:15:52.240 }, 00:15:52.240 { 00:15:52.240 "dma_device_id": "system", 00:15:52.240 "dma_device_type": 1 00:15:52.240 }, 00:15:52.240 { 00:15:52.240 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:52.240 "dma_device_type": 2 00:15:52.240 } 00:15:52.240 ], 00:15:52.240 "driver_specific": { 00:15:52.240 "raid": { 00:15:52.240 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:52.240 "strip_size_kb": 0, 00:15:52.240 "state": "online", 00:15:52.240 "raid_level": "raid1", 00:15:52.240 "superblock": true, 00:15:52.240 "num_base_bdevs": 2, 00:15:52.240 "num_base_bdevs_discovered": 2, 00:15:52.240 "num_base_bdevs_operational": 2, 00:15:52.240 "base_bdevs_list": [ 00:15:52.240 { 00:15:52.240 "name": "pt1", 00:15:52.240 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:52.240 "is_configured": true, 00:15:52.240 "data_offset": 256, 00:15:52.240 "data_size": 7936 00:15:52.240 }, 00:15:52.240 { 00:15:52.240 "name": "pt2", 00:15:52.240 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:52.240 "is_configured": true, 00:15:52.240 "data_offset": 256, 00:15:52.240 "data_size": 7936 00:15:52.240 } 00:15:52.240 ] 00:15:52.240 } 00:15:52.240 } 00:15:52.240 }' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:52.240 pt2' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:52.240 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.240 [2024-09-29 16:54:13.901069] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' 41d985d4-7f8d-4d10-8f0b-1fe972d845a7 '!=' 41d985d4-7f8d-4d10-8f0b-1fe972d845a7 ']' 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.500 [2024-09-29 16:54:13.944802] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.500 16:54:13 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.500 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:52.500 "name": "raid_bdev1", 00:15:52.500 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:52.500 "strip_size_kb": 0, 00:15:52.500 "state": "online", 00:15:52.500 "raid_level": "raid1", 00:15:52.500 "superblock": true, 00:15:52.500 "num_base_bdevs": 2, 00:15:52.500 "num_base_bdevs_discovered": 1, 00:15:52.500 "num_base_bdevs_operational": 1, 00:15:52.500 "base_bdevs_list": [ 00:15:52.500 { 00:15:52.500 "name": null, 00:15:52.500 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:52.500 "is_configured": false, 00:15:52.500 "data_offset": 0, 00:15:52.500 "data_size": 7936 00:15:52.500 }, 00:15:52.500 { 00:15:52.500 "name": "pt2", 00:15:52.500 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:52.500 "is_configured": true, 00:15:52.500 "data_offset": 256, 00:15:52.500 "data_size": 7936 00:15:52.500 } 00:15:52.500 ] 00:15:52.500 }' 00:15:52.500 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:52.500 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.760 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:52.760 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.760 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.760 [2024-09-29 16:54:14.380002] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:52.761 [2024-09-29 16:54:14.380029] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:52.761 [2024-09-29 16:54:14.380076] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:52.761 [2024-09-29 16:54:14.380111] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:52.761 [2024-09-29 16:54:14.380120] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:52.761 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.761 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:52.761 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:52.761 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.761 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:52.761 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.021 [2024-09-29 16:54:14.455864] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:53.021 [2024-09-29 16:54:14.455909] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.021 [2024-09-29 16:54:14.455927] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:15:53.021 [2024-09-29 16:54:14.455935] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.021 [2024-09-29 16:54:14.457857] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.021 [2024-09-29 16:54:14.457887] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:53.021 [2024-09-29 16:54:14.457929] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:53.021 [2024-09-29 16:54:14.457968] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:53.021 [2024-09-29 16:54:14.458030] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:53.021 [2024-09-29 16:54:14.458038] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:53.021 [2024-09-29 16:54:14.458131] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:53.021 [2024-09-29 16:54:14.458206] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:53.021 [2024-09-29 16:54:14.458239] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:53.021 [2024-09-29 16:54:14.458309] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:53.021 pt2 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.021 "name": "raid_bdev1", 00:15:53.021 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:53.021 "strip_size_kb": 0, 00:15:53.021 "state": "online", 00:15:53.021 "raid_level": "raid1", 00:15:53.021 "superblock": true, 00:15:53.021 "num_base_bdevs": 2, 00:15:53.021 "num_base_bdevs_discovered": 1, 00:15:53.021 "num_base_bdevs_operational": 1, 00:15:53.021 "base_bdevs_list": [ 00:15:53.021 { 00:15:53.021 "name": null, 00:15:53.021 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:53.021 "is_configured": false, 00:15:53.021 "data_offset": 256, 00:15:53.021 "data_size": 7936 00:15:53.021 }, 00:15:53.021 { 00:15:53.021 "name": "pt2", 00:15:53.021 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:53.021 "is_configured": true, 00:15:53.021 "data_offset": 256, 00:15:53.021 "data_size": 7936 00:15:53.021 } 00:15:53.021 ] 00:15:53.021 }' 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.021 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.281 [2024-09-29 16:54:14.895149] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:53.281 [2024-09-29 16:54:14.895172] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:53.281 [2024-09-29 16:54:14.895224] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:53.281 [2024-09-29 16:54:14.895257] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:53.281 [2024-09-29 16:54:14.895267] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.281 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.541 [2024-09-29 16:54:14.955069] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:53.541 [2024-09-29 16:54:14.955115] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:53.541 [2024-09-29 16:54:14.955133] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:15:53.541 [2024-09-29 16:54:14.955145] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:53.541 [2024-09-29 16:54:14.957158] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:53.541 [2024-09-29 16:54:14.957203] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:53.541 [2024-09-29 16:54:14.957243] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:53.541 [2024-09-29 16:54:14.957272] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:53.541 [2024-09-29 16:54:14.957365] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:53.541 [2024-09-29 16:54:14.957385] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:53.541 [2024-09-29 16:54:14.957436] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:53.541 [2024-09-29 16:54:14.957498] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:53.541 [2024-09-29 16:54:14.957556] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:53.541 [2024-09-29 16:54:14.957566] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:53.541 [2024-09-29 16:54:14.957623] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:53.541 [2024-09-29 16:54:14.957710] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:53.541 [2024-09-29 16:54:14.957751] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:53.541 [2024-09-29 16:54:14.957826] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:53.541 pt1 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.541 "name": "raid_bdev1", 00:15:53.541 "uuid": "41d985d4-7f8d-4d10-8f0b-1fe972d845a7", 00:15:53.541 "strip_size_kb": 0, 00:15:53.541 "state": "online", 00:15:53.541 "raid_level": "raid1", 00:15:53.541 "superblock": true, 00:15:53.541 "num_base_bdevs": 2, 00:15:53.541 "num_base_bdevs_discovered": 1, 00:15:53.541 "num_base_bdevs_operational": 1, 00:15:53.541 "base_bdevs_list": [ 00:15:53.541 { 00:15:53.541 "name": null, 00:15:53.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:53.541 "is_configured": false, 00:15:53.541 "data_offset": 256, 00:15:53.541 "data_size": 7936 00:15:53.541 }, 00:15:53.541 { 00:15:53.541 "name": "pt2", 00:15:53.541 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:53.541 "is_configured": true, 00:15:53.541 "data_offset": 256, 00:15:53.541 "data_size": 7936 00:15:53.541 } 00:15:53.541 ] 00:15:53.541 }' 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.541 16:54:14 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:53.801 [2024-09-29 16:54:15.430551] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' 41d985d4-7f8d-4d10-8f0b-1fe972d845a7 '!=' 41d985d4-7f8d-4d10-8f0b-1fe972d845a7 ']' 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 97425 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@950 -- # '[' -z 97425 ']' 00:15:53.801 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # kill -0 97425 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # uname 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97425 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:54.062 killing process with pid 97425 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97425' 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@969 -- # kill 97425 00:15:54.062 [2024-09-29 16:54:15.513273] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:54.062 [2024-09-29 16:54:15.513329] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:54.062 [2024-09-29 16:54:15.513392] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:54.062 [2024-09-29 16:54:15.513406] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:54.062 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@974 -- # wait 97425 00:15:54.062 [2024-09-29 16:54:15.537712] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:54.322 16:54:15 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:15:54.322 00:15:54.322 real 0m4.897s 00:15:54.322 user 0m7.950s 00:15:54.322 sys 0m1.079s 00:15:54.322 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:54.322 16:54:15 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.322 ************************************ 00:15:54.322 END TEST raid_superblock_test_md_separate 00:15:54.322 ************************************ 00:15:54.322 16:54:15 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:15:54.322 16:54:15 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:15:54.322 16:54:15 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:15:54.322 16:54:15 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:54.322 16:54:15 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:54.322 ************************************ 00:15:54.322 START TEST raid_rebuild_test_sb_md_separate 00:15:54.322 ************************************ 00:15:54.322 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:15:54.322 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:15:54.322 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:15:54.322 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:54.322 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=97732 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 97732 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 97732 ']' 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:54.323 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:54.323 16:54:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:54.323 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:54.323 Zero copy mechanism will not be used. 00:15:54.323 [2024-09-29 16:54:15.961003] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:15:54.323 [2024-09-29 16:54:15.961115] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97732 ] 00:15:54.583 [2024-09-29 16:54:16.104888] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:54.583 [2024-09-29 16:54:16.151204] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:54.583 [2024-09-29 16:54:16.193842] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:54.583 [2024-09-29 16:54:16.193882] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.152 BaseBdev1_malloc 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.152 [2024-09-29 16:54:16.808866] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:55.152 [2024-09-29 16:54:16.808921] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:55.152 [2024-09-29 16:54:16.808943] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:55.152 [2024-09-29 16:54:16.808952] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:55.152 [2024-09-29 16:54:16.810845] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:55.152 [2024-09-29 16:54:16.810878] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:55.152 BaseBdev1 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.152 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.412 BaseBdev2_malloc 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.412 [2024-09-29 16:54:16.853686] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:55.412 [2024-09-29 16:54:16.853792] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:55.412 [2024-09-29 16:54:16.853836] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:55.412 [2024-09-29 16:54:16.853856] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:55.412 [2024-09-29 16:54:16.857614] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:55.412 [2024-09-29 16:54:16.857658] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:55.412 BaseBdev2 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.412 spare_malloc 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.412 spare_delay 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.412 [2024-09-29 16:54:16.896392] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:55.412 [2024-09-29 16:54:16.896438] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:55.412 [2024-09-29 16:54:16.896458] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:15:55.412 [2024-09-29 16:54:16.896469] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:55.412 [2024-09-29 16:54:16.898385] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:55.412 [2024-09-29 16:54:16.898415] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:55.412 spare 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.412 [2024-09-29 16:54:16.908440] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:55.412 [2024-09-29 16:54:16.910210] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:55.412 [2024-09-29 16:54:16.910357] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:55.412 [2024-09-29 16:54:16.910369] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:15:55.412 [2024-09-29 16:54:16.910442] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:15:55.412 [2024-09-29 16:54:16.910555] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:55.412 [2024-09-29 16:54:16.910570] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:55.412 [2024-09-29 16:54:16.910647] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.412 "name": "raid_bdev1", 00:15:55.412 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:15:55.412 "strip_size_kb": 0, 00:15:55.412 "state": "online", 00:15:55.412 "raid_level": "raid1", 00:15:55.412 "superblock": true, 00:15:55.412 "num_base_bdevs": 2, 00:15:55.412 "num_base_bdevs_discovered": 2, 00:15:55.412 "num_base_bdevs_operational": 2, 00:15:55.412 "base_bdevs_list": [ 00:15:55.412 { 00:15:55.412 "name": "BaseBdev1", 00:15:55.412 "uuid": "7a8e57a8-f09e-5bfb-8dc6-2d962b706628", 00:15:55.412 "is_configured": true, 00:15:55.412 "data_offset": 256, 00:15:55.412 "data_size": 7936 00:15:55.412 }, 00:15:55.412 { 00:15:55.412 "name": "BaseBdev2", 00:15:55.412 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:15:55.412 "is_configured": true, 00:15:55.412 "data_offset": 256, 00:15:55.412 "data_size": 7936 00:15:55.412 } 00:15:55.412 ] 00:15:55.412 }' 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.412 16:54:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.981 [2024-09-29 16:54:17.351910] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:55.981 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:55.982 [2024-09-29 16:54:17.611368] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:55.982 /dev/nbd0 00:15:55.982 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:56.241 1+0 records in 00:15:56.241 1+0 records out 00:15:56.241 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000359831 s, 11.4 MB/s 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:15:56.241 16:54:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:15:56.810 7936+0 records in 00:15:56.810 7936+0 records out 00:15:56.810 32505856 bytes (33 MB, 31 MiB) copied, 0.605321 s, 53.7 MB/s 00:15:56.810 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:56.810 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:56.810 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:56.810 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:56.810 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:15:56.810 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:56.810 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:57.070 [2024-09-29 16:54:18.491729] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.070 [2024-09-29 16:54:18.524266] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.070 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:57.070 "name": "raid_bdev1", 00:15:57.070 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:15:57.070 "strip_size_kb": 0, 00:15:57.070 "state": "online", 00:15:57.070 "raid_level": "raid1", 00:15:57.070 "superblock": true, 00:15:57.070 "num_base_bdevs": 2, 00:15:57.070 "num_base_bdevs_discovered": 1, 00:15:57.070 "num_base_bdevs_operational": 1, 00:15:57.070 "base_bdevs_list": [ 00:15:57.070 { 00:15:57.071 "name": null, 00:15:57.071 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.071 "is_configured": false, 00:15:57.071 "data_offset": 0, 00:15:57.071 "data_size": 7936 00:15:57.071 }, 00:15:57.071 { 00:15:57.071 "name": "BaseBdev2", 00:15:57.071 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:15:57.071 "is_configured": true, 00:15:57.071 "data_offset": 256, 00:15:57.071 "data_size": 7936 00:15:57.071 } 00:15:57.071 ] 00:15:57.071 }' 00:15:57.071 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:57.071 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.330 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:57.330 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.330 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:57.330 [2024-09-29 16:54:18.967504] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:57.330 [2024-09-29 16:54:18.969318] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:15:57.330 [2024-09-29 16:54:18.971189] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:57.330 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.330 16:54:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.712 16:54:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:58.712 "name": "raid_bdev1", 00:15:58.712 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:15:58.712 "strip_size_kb": 0, 00:15:58.712 "state": "online", 00:15:58.712 "raid_level": "raid1", 00:15:58.712 "superblock": true, 00:15:58.712 "num_base_bdevs": 2, 00:15:58.712 "num_base_bdevs_discovered": 2, 00:15:58.712 "num_base_bdevs_operational": 2, 00:15:58.712 "process": { 00:15:58.712 "type": "rebuild", 00:15:58.712 "target": "spare", 00:15:58.712 "progress": { 00:15:58.712 "blocks": 2560, 00:15:58.712 "percent": 32 00:15:58.712 } 00:15:58.712 }, 00:15:58.712 "base_bdevs_list": [ 00:15:58.712 { 00:15:58.712 "name": "spare", 00:15:58.712 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:15:58.712 "is_configured": true, 00:15:58.712 "data_offset": 256, 00:15:58.712 "data_size": 7936 00:15:58.712 }, 00:15:58.712 { 00:15:58.712 "name": "BaseBdev2", 00:15:58.712 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:15:58.712 "is_configured": true, 00:15:58.712 "data_offset": 256, 00:15:58.712 "data_size": 7936 00:15:58.712 } 00:15:58.712 ] 00:15:58.712 }' 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.712 [2024-09-29 16:54:20.134467] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:58.712 [2024-09-29 16:54:20.175966] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:58.712 [2024-09-29 16:54:20.176019] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:58.712 [2024-09-29 16:54:20.176036] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:58.712 [2024-09-29 16:54:20.176052] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.712 "name": "raid_bdev1", 00:15:58.712 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:15:58.712 "strip_size_kb": 0, 00:15:58.712 "state": "online", 00:15:58.712 "raid_level": "raid1", 00:15:58.712 "superblock": true, 00:15:58.712 "num_base_bdevs": 2, 00:15:58.712 "num_base_bdevs_discovered": 1, 00:15:58.712 "num_base_bdevs_operational": 1, 00:15:58.712 "base_bdevs_list": [ 00:15:58.712 { 00:15:58.712 "name": null, 00:15:58.712 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:58.712 "is_configured": false, 00:15:58.712 "data_offset": 0, 00:15:58.712 "data_size": 7936 00:15:58.712 }, 00:15:58.712 { 00:15:58.712 "name": "BaseBdev2", 00:15:58.712 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:15:58.712 "is_configured": true, 00:15:58.712 "data_offset": 256, 00:15:58.712 "data_size": 7936 00:15:58.712 } 00:15:58.712 ] 00:15:58.712 }' 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.712 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:58.972 "name": "raid_bdev1", 00:15:58.972 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:15:58.972 "strip_size_kb": 0, 00:15:58.972 "state": "online", 00:15:58.972 "raid_level": "raid1", 00:15:58.972 "superblock": true, 00:15:58.972 "num_base_bdevs": 2, 00:15:58.972 "num_base_bdevs_discovered": 1, 00:15:58.972 "num_base_bdevs_operational": 1, 00:15:58.972 "base_bdevs_list": [ 00:15:58.972 { 00:15:58.972 "name": null, 00:15:58.972 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:58.972 "is_configured": false, 00:15:58.972 "data_offset": 0, 00:15:58.972 "data_size": 7936 00:15:58.972 }, 00:15:58.972 { 00:15:58.972 "name": "BaseBdev2", 00:15:58.972 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:15:58.972 "is_configured": true, 00:15:58.972 "data_offset": 256, 00:15:58.972 "data_size": 7936 00:15:58.972 } 00:15:58.972 ] 00:15:58.972 }' 00:15:58.972 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:15:59.232 [2024-09-29 16:54:20.726129] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:59.232 [2024-09-29 16:54:20.727691] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:15:59.232 [2024-09-29 16:54:20.729527] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.232 16:54:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:00.171 "name": "raid_bdev1", 00:16:00.171 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:00.171 "strip_size_kb": 0, 00:16:00.171 "state": "online", 00:16:00.171 "raid_level": "raid1", 00:16:00.171 "superblock": true, 00:16:00.171 "num_base_bdevs": 2, 00:16:00.171 "num_base_bdevs_discovered": 2, 00:16:00.171 "num_base_bdevs_operational": 2, 00:16:00.171 "process": { 00:16:00.171 "type": "rebuild", 00:16:00.171 "target": "spare", 00:16:00.171 "progress": { 00:16:00.171 "blocks": 2560, 00:16:00.171 "percent": 32 00:16:00.171 } 00:16:00.171 }, 00:16:00.171 "base_bdevs_list": [ 00:16:00.171 { 00:16:00.171 "name": "spare", 00:16:00.171 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:00.171 "is_configured": true, 00:16:00.171 "data_offset": 256, 00:16:00.171 "data_size": 7936 00:16:00.171 }, 00:16:00.171 { 00:16:00.171 "name": "BaseBdev2", 00:16:00.171 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:00.171 "is_configured": true, 00:16:00.171 "data_offset": 256, 00:16:00.171 "data_size": 7936 00:16:00.171 } 00:16:00.171 ] 00:16:00.171 }' 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:00.171 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:00.432 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=586 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:00.432 "name": "raid_bdev1", 00:16:00.432 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:00.432 "strip_size_kb": 0, 00:16:00.432 "state": "online", 00:16:00.432 "raid_level": "raid1", 00:16:00.432 "superblock": true, 00:16:00.432 "num_base_bdevs": 2, 00:16:00.432 "num_base_bdevs_discovered": 2, 00:16:00.432 "num_base_bdevs_operational": 2, 00:16:00.432 "process": { 00:16:00.432 "type": "rebuild", 00:16:00.432 "target": "spare", 00:16:00.432 "progress": { 00:16:00.432 "blocks": 2816, 00:16:00.432 "percent": 35 00:16:00.432 } 00:16:00.432 }, 00:16:00.432 "base_bdevs_list": [ 00:16:00.432 { 00:16:00.432 "name": "spare", 00:16:00.432 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:00.432 "is_configured": true, 00:16:00.432 "data_offset": 256, 00:16:00.432 "data_size": 7936 00:16:00.432 }, 00:16:00.432 { 00:16:00.432 "name": "BaseBdev2", 00:16:00.432 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:00.432 "is_configured": true, 00:16:00.432 "data_offset": 256, 00:16:00.432 "data_size": 7936 00:16:00.432 } 00:16:00.432 ] 00:16:00.432 }' 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:00.432 16:54:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:00.432 16:54:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:00.432 16:54:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:01.814 "name": "raid_bdev1", 00:16:01.814 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:01.814 "strip_size_kb": 0, 00:16:01.814 "state": "online", 00:16:01.814 "raid_level": "raid1", 00:16:01.814 "superblock": true, 00:16:01.814 "num_base_bdevs": 2, 00:16:01.814 "num_base_bdevs_discovered": 2, 00:16:01.814 "num_base_bdevs_operational": 2, 00:16:01.814 "process": { 00:16:01.814 "type": "rebuild", 00:16:01.814 "target": "spare", 00:16:01.814 "progress": { 00:16:01.814 "blocks": 5888, 00:16:01.814 "percent": 74 00:16:01.814 } 00:16:01.814 }, 00:16:01.814 "base_bdevs_list": [ 00:16:01.814 { 00:16:01.814 "name": "spare", 00:16:01.814 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:01.814 "is_configured": true, 00:16:01.814 "data_offset": 256, 00:16:01.814 "data_size": 7936 00:16:01.814 }, 00:16:01.814 { 00:16:01.814 "name": "BaseBdev2", 00:16:01.814 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:01.814 "is_configured": true, 00:16:01.814 "data_offset": 256, 00:16:01.814 "data_size": 7936 00:16:01.814 } 00:16:01.814 ] 00:16:01.814 }' 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:01.814 16:54:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:02.383 [2024-09-29 16:54:23.840091] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:02.383 [2024-09-29 16:54:23.840169] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:02.383 [2024-09-29 16:54:23.840272] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:02.643 "name": "raid_bdev1", 00:16:02.643 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:02.643 "strip_size_kb": 0, 00:16:02.643 "state": "online", 00:16:02.643 "raid_level": "raid1", 00:16:02.643 "superblock": true, 00:16:02.643 "num_base_bdevs": 2, 00:16:02.643 "num_base_bdevs_discovered": 2, 00:16:02.643 "num_base_bdevs_operational": 2, 00:16:02.643 "base_bdevs_list": [ 00:16:02.643 { 00:16:02.643 "name": "spare", 00:16:02.643 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:02.643 "is_configured": true, 00:16:02.643 "data_offset": 256, 00:16:02.643 "data_size": 7936 00:16:02.643 }, 00:16:02.643 { 00:16:02.643 "name": "BaseBdev2", 00:16:02.643 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:02.643 "is_configured": true, 00:16:02.643 "data_offset": 256, 00:16:02.643 "data_size": 7936 00:16:02.643 } 00:16:02.643 ] 00:16:02.643 }' 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:02.643 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:02.902 "name": "raid_bdev1", 00:16:02.902 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:02.902 "strip_size_kb": 0, 00:16:02.902 "state": "online", 00:16:02.902 "raid_level": "raid1", 00:16:02.902 "superblock": true, 00:16:02.902 "num_base_bdevs": 2, 00:16:02.902 "num_base_bdevs_discovered": 2, 00:16:02.902 "num_base_bdevs_operational": 2, 00:16:02.902 "base_bdevs_list": [ 00:16:02.902 { 00:16:02.902 "name": "spare", 00:16:02.902 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:02.902 "is_configured": true, 00:16:02.902 "data_offset": 256, 00:16:02.902 "data_size": 7936 00:16:02.902 }, 00:16:02.902 { 00:16:02.902 "name": "BaseBdev2", 00:16:02.902 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:02.902 "is_configured": true, 00:16:02.902 "data_offset": 256, 00:16:02.902 "data_size": 7936 00:16:02.902 } 00:16:02.902 ] 00:16:02.902 }' 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.902 "name": "raid_bdev1", 00:16:02.902 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:02.902 "strip_size_kb": 0, 00:16:02.902 "state": "online", 00:16:02.902 "raid_level": "raid1", 00:16:02.902 "superblock": true, 00:16:02.902 "num_base_bdevs": 2, 00:16:02.902 "num_base_bdevs_discovered": 2, 00:16:02.902 "num_base_bdevs_operational": 2, 00:16:02.902 "base_bdevs_list": [ 00:16:02.902 { 00:16:02.902 "name": "spare", 00:16:02.902 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:02.902 "is_configured": true, 00:16:02.902 "data_offset": 256, 00:16:02.902 "data_size": 7936 00:16:02.902 }, 00:16:02.902 { 00:16:02.902 "name": "BaseBdev2", 00:16:02.902 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:02.902 "is_configured": true, 00:16:02.902 "data_offset": 256, 00:16:02.902 "data_size": 7936 00:16:02.902 } 00:16:02.902 ] 00:16:02.902 }' 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.902 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.472 [2024-09-29 16:54:24.877164] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:03.472 [2024-09-29 16:54:24.877194] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:03.472 [2024-09-29 16:54:24.877272] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:03.472 [2024-09-29 16:54:24.877334] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:03.472 [2024-09-29 16:54:24.877346] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:03.472 16:54:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:03.472 /dev/nbd0 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:03.731 1+0 records in 00:16:03.731 1+0 records out 00:16:03.731 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000482263 s, 8.5 MB/s 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:03.731 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:03.732 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:16:03.732 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:03.732 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:03.732 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:03.732 /dev/nbd1 00:16:03.991 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:03.991 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:03.991 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:03.992 1+0 records in 00:16:03.992 1+0 records out 00:16:03.992 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000351809 s, 11.6 MB/s 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:03.992 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:04.252 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.253 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.512 [2024-09-29 16:54:25.929791] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:04.512 [2024-09-29 16:54:25.929838] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.512 [2024-09-29 16:54:25.929856] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:16:04.512 [2024-09-29 16:54:25.929877] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.512 [2024-09-29 16:54:25.931773] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.512 [2024-09-29 16:54:25.931819] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:04.513 [2024-09-29 16:54:25.931869] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:04.513 [2024-09-29 16:54:25.931916] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:04.513 [2024-09-29 16:54:25.932039] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:04.513 spare 00:16:04.513 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.513 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:04.513 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.513 16:54:25 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.513 [2024-09-29 16:54:26.031947] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:04.513 [2024-09-29 16:54:26.031973] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:04.513 [2024-09-29 16:54:26.032088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:16:04.513 [2024-09-29 16:54:26.032179] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:04.513 [2024-09-29 16:54:26.032190] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:04.513 [2024-09-29 16:54:26.032277] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.513 "name": "raid_bdev1", 00:16:04.513 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:04.513 "strip_size_kb": 0, 00:16:04.513 "state": "online", 00:16:04.513 "raid_level": "raid1", 00:16:04.513 "superblock": true, 00:16:04.513 "num_base_bdevs": 2, 00:16:04.513 "num_base_bdevs_discovered": 2, 00:16:04.513 "num_base_bdevs_operational": 2, 00:16:04.513 "base_bdevs_list": [ 00:16:04.513 { 00:16:04.513 "name": "spare", 00:16:04.513 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:04.513 "is_configured": true, 00:16:04.513 "data_offset": 256, 00:16:04.513 "data_size": 7936 00:16:04.513 }, 00:16:04.513 { 00:16:04.513 "name": "BaseBdev2", 00:16:04.513 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:04.513 "is_configured": true, 00:16:04.513 "data_offset": 256, 00:16:04.513 "data_size": 7936 00:16:04.513 } 00:16:04.513 ] 00:16:04.513 }' 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.513 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:05.083 "name": "raid_bdev1", 00:16:05.083 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:05.083 "strip_size_kb": 0, 00:16:05.083 "state": "online", 00:16:05.083 "raid_level": "raid1", 00:16:05.083 "superblock": true, 00:16:05.083 "num_base_bdevs": 2, 00:16:05.083 "num_base_bdevs_discovered": 2, 00:16:05.083 "num_base_bdevs_operational": 2, 00:16:05.083 "base_bdevs_list": [ 00:16:05.083 { 00:16:05.083 "name": "spare", 00:16:05.083 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:05.083 "is_configured": true, 00:16:05.083 "data_offset": 256, 00:16:05.083 "data_size": 7936 00:16:05.083 }, 00:16:05.083 { 00:16:05.083 "name": "BaseBdev2", 00:16:05.083 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:05.083 "is_configured": true, 00:16:05.083 "data_offset": 256, 00:16:05.083 "data_size": 7936 00:16:05.083 } 00:16:05.083 ] 00:16:05.083 }' 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.083 [2024-09-29 16:54:26.664601] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:05.083 "name": "raid_bdev1", 00:16:05.083 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:05.083 "strip_size_kb": 0, 00:16:05.083 "state": "online", 00:16:05.083 "raid_level": "raid1", 00:16:05.083 "superblock": true, 00:16:05.083 "num_base_bdevs": 2, 00:16:05.083 "num_base_bdevs_discovered": 1, 00:16:05.083 "num_base_bdevs_operational": 1, 00:16:05.083 "base_bdevs_list": [ 00:16:05.083 { 00:16:05.083 "name": null, 00:16:05.083 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:05.083 "is_configured": false, 00:16:05.083 "data_offset": 0, 00:16:05.083 "data_size": 7936 00:16:05.083 }, 00:16:05.083 { 00:16:05.083 "name": "BaseBdev2", 00:16:05.083 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:05.083 "is_configured": true, 00:16:05.083 "data_offset": 256, 00:16:05.083 "data_size": 7936 00:16:05.083 } 00:16:05.083 ] 00:16:05.083 }' 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:05.083 16:54:26 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.653 16:54:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:05.653 16:54:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.653 16:54:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:05.653 [2024-09-29 16:54:27.063914] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:05.653 [2024-09-29 16:54:27.064059] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:05.653 [2024-09-29 16:54:27.064074] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:05.653 [2024-09-29 16:54:27.064107] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:05.653 [2024-09-29 16:54:27.065830] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:16:05.653 [2024-09-29 16:54:27.067690] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:05.653 16:54:27 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.653 16:54:27 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:06.591 "name": "raid_bdev1", 00:16:06.591 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:06.591 "strip_size_kb": 0, 00:16:06.591 "state": "online", 00:16:06.591 "raid_level": "raid1", 00:16:06.591 "superblock": true, 00:16:06.591 "num_base_bdevs": 2, 00:16:06.591 "num_base_bdevs_discovered": 2, 00:16:06.591 "num_base_bdevs_operational": 2, 00:16:06.591 "process": { 00:16:06.591 "type": "rebuild", 00:16:06.591 "target": "spare", 00:16:06.591 "progress": { 00:16:06.591 "blocks": 2560, 00:16:06.591 "percent": 32 00:16:06.591 } 00:16:06.591 }, 00:16:06.591 "base_bdevs_list": [ 00:16:06.591 { 00:16:06.591 "name": "spare", 00:16:06.591 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:06.591 "is_configured": true, 00:16:06.591 "data_offset": 256, 00:16:06.591 "data_size": 7936 00:16:06.591 }, 00:16:06.591 { 00:16:06.591 "name": "BaseBdev2", 00:16:06.591 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:06.591 "is_configured": true, 00:16:06.591 "data_offset": 256, 00:16:06.591 "data_size": 7936 00:16:06.591 } 00:16:06.591 ] 00:16:06.591 }' 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.591 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.591 [2024-09-29 16:54:28.207015] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:06.851 [2024-09-29 16:54:28.271815] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:06.851 [2024-09-29 16:54:28.271863] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.851 [2024-09-29 16:54:28.271880] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:06.851 [2024-09-29 16:54:28.271887] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.851 "name": "raid_bdev1", 00:16:06.851 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:06.851 "strip_size_kb": 0, 00:16:06.851 "state": "online", 00:16:06.851 "raid_level": "raid1", 00:16:06.851 "superblock": true, 00:16:06.851 "num_base_bdevs": 2, 00:16:06.851 "num_base_bdevs_discovered": 1, 00:16:06.851 "num_base_bdevs_operational": 1, 00:16:06.851 "base_bdevs_list": [ 00:16:06.851 { 00:16:06.851 "name": null, 00:16:06.851 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.851 "is_configured": false, 00:16:06.851 "data_offset": 0, 00:16:06.851 "data_size": 7936 00:16:06.851 }, 00:16:06.851 { 00:16:06.851 "name": "BaseBdev2", 00:16:06.851 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:06.851 "is_configured": true, 00:16:06.851 "data_offset": 256, 00:16:06.851 "data_size": 7936 00:16:06.851 } 00:16:06.851 ] 00:16:06.851 }' 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.851 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:07.110 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:07.110 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.110 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:07.110 [2024-09-29 16:54:28.730151] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:07.110 [2024-09-29 16:54:28.730203] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:07.110 [2024-09-29 16:54:28.730227] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:16:07.110 [2024-09-29 16:54:28.730235] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:07.110 [2024-09-29 16:54:28.730432] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:07.110 [2024-09-29 16:54:28.730454] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:07.110 [2024-09-29 16:54:28.730506] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:07.110 [2024-09-29 16:54:28.730515] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:07.110 [2024-09-29 16:54:28.730532] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:07.110 [2024-09-29 16:54:28.730549] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:07.110 [2024-09-29 16:54:28.731997] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:16:07.110 [2024-09-29 16:54:28.733805] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:07.110 spare 00:16:07.110 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.110 16:54:28 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:08.492 "name": "raid_bdev1", 00:16:08.492 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:08.492 "strip_size_kb": 0, 00:16:08.492 "state": "online", 00:16:08.492 "raid_level": "raid1", 00:16:08.492 "superblock": true, 00:16:08.492 "num_base_bdevs": 2, 00:16:08.492 "num_base_bdevs_discovered": 2, 00:16:08.492 "num_base_bdevs_operational": 2, 00:16:08.492 "process": { 00:16:08.492 "type": "rebuild", 00:16:08.492 "target": "spare", 00:16:08.492 "progress": { 00:16:08.492 "blocks": 2560, 00:16:08.492 "percent": 32 00:16:08.492 } 00:16:08.492 }, 00:16:08.492 "base_bdevs_list": [ 00:16:08.492 { 00:16:08.492 "name": "spare", 00:16:08.492 "uuid": "3b54ac0b-7e1f-505d-938f-3b0d5269329e", 00:16:08.492 "is_configured": true, 00:16:08.492 "data_offset": 256, 00:16:08.492 "data_size": 7936 00:16:08.492 }, 00:16:08.492 { 00:16:08.492 "name": "BaseBdev2", 00:16:08.492 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:08.492 "is_configured": true, 00:16:08.492 "data_offset": 256, 00:16:08.492 "data_size": 7936 00:16:08.492 } 00:16:08.492 ] 00:16:08.492 }' 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.492 [2024-09-29 16:54:29.897132] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.492 [2024-09-29 16:54:29.937919] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:08.492 [2024-09-29 16:54:29.937975] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:08.492 [2024-09-29 16:54:29.937989] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:08.492 [2024-09-29 16:54:29.937997] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:08.492 "name": "raid_bdev1", 00:16:08.492 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:08.492 "strip_size_kb": 0, 00:16:08.492 "state": "online", 00:16:08.492 "raid_level": "raid1", 00:16:08.492 "superblock": true, 00:16:08.492 "num_base_bdevs": 2, 00:16:08.492 "num_base_bdevs_discovered": 1, 00:16:08.492 "num_base_bdevs_operational": 1, 00:16:08.492 "base_bdevs_list": [ 00:16:08.492 { 00:16:08.492 "name": null, 00:16:08.492 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.492 "is_configured": false, 00:16:08.492 "data_offset": 0, 00:16:08.492 "data_size": 7936 00:16:08.492 }, 00:16:08.492 { 00:16:08.492 "name": "BaseBdev2", 00:16:08.492 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:08.492 "is_configured": true, 00:16:08.492 "data_offset": 256, 00:16:08.492 "data_size": 7936 00:16:08.492 } 00:16:08.492 ] 00:16:08.492 }' 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:08.492 16:54:29 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:08.752 "name": "raid_bdev1", 00:16:08.752 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:08.752 "strip_size_kb": 0, 00:16:08.752 "state": "online", 00:16:08.752 "raid_level": "raid1", 00:16:08.752 "superblock": true, 00:16:08.752 "num_base_bdevs": 2, 00:16:08.752 "num_base_bdevs_discovered": 1, 00:16:08.752 "num_base_bdevs_operational": 1, 00:16:08.752 "base_bdevs_list": [ 00:16:08.752 { 00:16:08.752 "name": null, 00:16:08.752 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:08.752 "is_configured": false, 00:16:08.752 "data_offset": 0, 00:16:08.752 "data_size": 7936 00:16:08.752 }, 00:16:08.752 { 00:16:08.752 "name": "BaseBdev2", 00:16:08.752 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:08.752 "is_configured": true, 00:16:08.752 "data_offset": 256, 00:16:08.752 "data_size": 7936 00:16:08.752 } 00:16:08.752 ] 00:16:08.752 }' 00:16:08.752 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:09.011 [2024-09-29 16:54:30.508077] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:09.011 [2024-09-29 16:54:30.508124] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:09.011 [2024-09-29 16:54:30.508141] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:16:09.011 [2024-09-29 16:54:30.508152] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:09.011 [2024-09-29 16:54:30.508325] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:09.011 [2024-09-29 16:54:30.508342] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:09.011 [2024-09-29 16:54:30.508384] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:09.011 [2024-09-29 16:54:30.508411] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:09.011 [2024-09-29 16:54:30.508424] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:09.011 [2024-09-29 16:54:30.508437] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:09.011 BaseBdev1 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.011 16:54:30 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.980 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:09.980 "name": "raid_bdev1", 00:16:09.980 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:09.980 "strip_size_kb": 0, 00:16:09.980 "state": "online", 00:16:09.980 "raid_level": "raid1", 00:16:09.980 "superblock": true, 00:16:09.980 "num_base_bdevs": 2, 00:16:09.980 "num_base_bdevs_discovered": 1, 00:16:09.980 "num_base_bdevs_operational": 1, 00:16:09.980 "base_bdevs_list": [ 00:16:09.980 { 00:16:09.980 "name": null, 00:16:09.980 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:09.980 "is_configured": false, 00:16:09.980 "data_offset": 0, 00:16:09.980 "data_size": 7936 00:16:09.980 }, 00:16:09.980 { 00:16:09.981 "name": "BaseBdev2", 00:16:09.981 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:09.981 "is_configured": true, 00:16:09.981 "data_offset": 256, 00:16:09.981 "data_size": 7936 00:16:09.981 } 00:16:09.981 ] 00:16:09.981 }' 00:16:09.981 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:09.981 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.265 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:10.265 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:10.265 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:10.265 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:10.265 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:10.265 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:10.265 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:10.525 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.525 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.525 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.525 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:10.525 "name": "raid_bdev1", 00:16:10.525 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:10.525 "strip_size_kb": 0, 00:16:10.525 "state": "online", 00:16:10.525 "raid_level": "raid1", 00:16:10.525 "superblock": true, 00:16:10.525 "num_base_bdevs": 2, 00:16:10.525 "num_base_bdevs_discovered": 1, 00:16:10.525 "num_base_bdevs_operational": 1, 00:16:10.525 "base_bdevs_list": [ 00:16:10.525 { 00:16:10.525 "name": null, 00:16:10.525 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:10.525 "is_configured": false, 00:16:10.525 "data_offset": 0, 00:16:10.525 "data_size": 7936 00:16:10.525 }, 00:16:10.525 { 00:16:10.525 "name": "BaseBdev2", 00:16:10.525 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:10.525 "is_configured": true, 00:16:10.525 "data_offset": 256, 00:16:10.525 "data_size": 7936 00:16:10.525 } 00:16:10.525 ] 00:16:10.525 }' 00:16:10.525 16:54:31 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:10.525 [2024-09-29 16:54:32.069437] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:10.525 [2024-09-29 16:54:32.069581] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:10.525 [2024-09-29 16:54:32.069593] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:10.525 request: 00:16:10.525 { 00:16:10.525 "base_bdev": "BaseBdev1", 00:16:10.525 "raid_bdev": "raid_bdev1", 00:16:10.525 "method": "bdev_raid_add_base_bdev", 00:16:10.525 "req_id": 1 00:16:10.525 } 00:16:10.525 Got JSON-RPC error response 00:16:10.525 response: 00:16:10.525 { 00:16:10.525 "code": -22, 00:16:10.525 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:10.525 } 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # es=1 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:10.525 16:54:32 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.464 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:11.464 "name": "raid_bdev1", 00:16:11.464 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:11.464 "strip_size_kb": 0, 00:16:11.464 "state": "online", 00:16:11.464 "raid_level": "raid1", 00:16:11.464 "superblock": true, 00:16:11.464 "num_base_bdevs": 2, 00:16:11.464 "num_base_bdevs_discovered": 1, 00:16:11.464 "num_base_bdevs_operational": 1, 00:16:11.464 "base_bdevs_list": [ 00:16:11.464 { 00:16:11.464 "name": null, 00:16:11.465 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:11.465 "is_configured": false, 00:16:11.465 "data_offset": 0, 00:16:11.465 "data_size": 7936 00:16:11.465 }, 00:16:11.465 { 00:16:11.465 "name": "BaseBdev2", 00:16:11.465 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:11.465 "is_configured": true, 00:16:11.465 "data_offset": 256, 00:16:11.465 "data_size": 7936 00:16:11.465 } 00:16:11.465 ] 00:16:11.465 }' 00:16:11.465 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:11.465 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:12.035 "name": "raid_bdev1", 00:16:12.035 "uuid": "0d53cb8a-8499-46ac-bc39-51d684f95e0d", 00:16:12.035 "strip_size_kb": 0, 00:16:12.035 "state": "online", 00:16:12.035 "raid_level": "raid1", 00:16:12.035 "superblock": true, 00:16:12.035 "num_base_bdevs": 2, 00:16:12.035 "num_base_bdevs_discovered": 1, 00:16:12.035 "num_base_bdevs_operational": 1, 00:16:12.035 "base_bdevs_list": [ 00:16:12.035 { 00:16:12.035 "name": null, 00:16:12.035 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:12.035 "is_configured": false, 00:16:12.035 "data_offset": 0, 00:16:12.035 "data_size": 7936 00:16:12.035 }, 00:16:12.035 { 00:16:12.035 "name": "BaseBdev2", 00:16:12.035 "uuid": "abeec679-2a4e-5235-9502-d8de40a277b8", 00:16:12.035 "is_configured": true, 00:16:12.035 "data_offset": 256, 00:16:12.035 "data_size": 7936 00:16:12.035 } 00:16:12.035 ] 00:16:12.035 }' 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 97732 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 97732 ']' 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 97732 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97732 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:12.035 killing process with pid 97732 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97732' 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 97732 00:16:12.035 Received shutdown signal, test time was about 60.000000 seconds 00:16:12.035 00:16:12.035 Latency(us) 00:16:12.035 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:12.035 =================================================================================================================== 00:16:12.035 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:12.035 [2024-09-29 16:54:33.667605] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:12.035 [2024-09-29 16:54:33.667750] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:12.035 [2024-09-29 16:54:33.667803] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:12.035 [2024-09-29 16:54:33.667813] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:12.035 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 97732 00:16:12.035 [2024-09-29 16:54:33.701204] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:12.296 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:16:12.296 00:16:12.296 real 0m18.064s 00:16:12.296 user 0m23.825s 00:16:12.296 sys 0m2.603s 00:16:12.296 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:12.296 16:54:33 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:12.296 ************************************ 00:16:12.296 END TEST raid_rebuild_test_sb_md_separate 00:16:12.296 ************************************ 00:16:12.556 16:54:33 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:16:12.556 16:54:33 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:16:12.556 16:54:33 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:16:12.556 16:54:33 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:12.556 16:54:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:12.556 ************************************ 00:16:12.556 START TEST raid_state_function_test_sb_md_interleaved 00:16:12.556 ************************************ 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=98407 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:12.556 Process raid pid: 98407 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98407' 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 98407 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 98407 ']' 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:12.556 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:12.556 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:12.557 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:12.557 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:12.557 [2024-09-29 16:54:34.109829] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:12.557 [2024-09-29 16:54:34.109962] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:12.817 [2024-09-29 16:54:34.256608] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:12.817 [2024-09-29 16:54:34.303109] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:12.817 [2024-09-29 16:54:34.346358] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:12.817 [2024-09-29 16:54:34.346395] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.388 [2024-09-29 16:54:34.912216] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:13.388 [2024-09-29 16:54:34.912263] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:13.388 [2024-09-29 16:54:34.912275] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:13.388 [2024-09-29 16:54:34.912284] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:13.388 "name": "Existed_Raid", 00:16:13.388 "uuid": "d1a0b283-f286-4c7a-a099-7a7d6c0dd188", 00:16:13.388 "strip_size_kb": 0, 00:16:13.388 "state": "configuring", 00:16:13.388 "raid_level": "raid1", 00:16:13.388 "superblock": true, 00:16:13.388 "num_base_bdevs": 2, 00:16:13.388 "num_base_bdevs_discovered": 0, 00:16:13.388 "num_base_bdevs_operational": 2, 00:16:13.388 "base_bdevs_list": [ 00:16:13.388 { 00:16:13.388 "name": "BaseBdev1", 00:16:13.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.388 "is_configured": false, 00:16:13.388 "data_offset": 0, 00:16:13.388 "data_size": 0 00:16:13.388 }, 00:16:13.388 { 00:16:13.388 "name": "BaseBdev2", 00:16:13.388 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.388 "is_configured": false, 00:16:13.388 "data_offset": 0, 00:16:13.388 "data_size": 0 00:16:13.388 } 00:16:13.388 ] 00:16:13.388 }' 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:13.388 16:54:34 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.959 [2024-09-29 16:54:35.383427] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:13.959 [2024-09-29 16:54:35.383475] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.959 [2024-09-29 16:54:35.395445] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:13.959 [2024-09-29 16:54:35.395481] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:13.959 [2024-09-29 16:54:35.395498] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:13.959 [2024-09-29 16:54:35.395509] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.959 [2024-09-29 16:54:35.416477] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:13.959 BaseBdev1 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.959 [ 00:16:13.959 { 00:16:13.959 "name": "BaseBdev1", 00:16:13.959 "aliases": [ 00:16:13.959 "57f6d7db-955a-4474-b78f-8b9ccf03e227" 00:16:13.959 ], 00:16:13.959 "product_name": "Malloc disk", 00:16:13.959 "block_size": 4128, 00:16:13.959 "num_blocks": 8192, 00:16:13.959 "uuid": "57f6d7db-955a-4474-b78f-8b9ccf03e227", 00:16:13.959 "md_size": 32, 00:16:13.959 "md_interleave": true, 00:16:13.959 "dif_type": 0, 00:16:13.959 "assigned_rate_limits": { 00:16:13.959 "rw_ios_per_sec": 0, 00:16:13.959 "rw_mbytes_per_sec": 0, 00:16:13.959 "r_mbytes_per_sec": 0, 00:16:13.959 "w_mbytes_per_sec": 0 00:16:13.959 }, 00:16:13.959 "claimed": true, 00:16:13.959 "claim_type": "exclusive_write", 00:16:13.959 "zoned": false, 00:16:13.959 "supported_io_types": { 00:16:13.959 "read": true, 00:16:13.959 "write": true, 00:16:13.959 "unmap": true, 00:16:13.959 "flush": true, 00:16:13.959 "reset": true, 00:16:13.959 "nvme_admin": false, 00:16:13.959 "nvme_io": false, 00:16:13.959 "nvme_io_md": false, 00:16:13.959 "write_zeroes": true, 00:16:13.959 "zcopy": true, 00:16:13.959 "get_zone_info": false, 00:16:13.959 "zone_management": false, 00:16:13.959 "zone_append": false, 00:16:13.959 "compare": false, 00:16:13.959 "compare_and_write": false, 00:16:13.959 "abort": true, 00:16:13.959 "seek_hole": false, 00:16:13.959 "seek_data": false, 00:16:13.959 "copy": true, 00:16:13.959 "nvme_iov_md": false 00:16:13.959 }, 00:16:13.959 "memory_domains": [ 00:16:13.959 { 00:16:13.959 "dma_device_id": "system", 00:16:13.959 "dma_device_type": 1 00:16:13.959 }, 00:16:13.959 { 00:16:13.959 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:13.959 "dma_device_type": 2 00:16:13.959 } 00:16:13.959 ], 00:16:13.959 "driver_specific": {} 00:16:13.959 } 00:16:13.959 ] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:13.959 "name": "Existed_Raid", 00:16:13.959 "uuid": "74d825d5-dab1-46d5-9d61-2d30b5ace3c1", 00:16:13.959 "strip_size_kb": 0, 00:16:13.959 "state": "configuring", 00:16:13.959 "raid_level": "raid1", 00:16:13.959 "superblock": true, 00:16:13.959 "num_base_bdevs": 2, 00:16:13.959 "num_base_bdevs_discovered": 1, 00:16:13.959 "num_base_bdevs_operational": 2, 00:16:13.959 "base_bdevs_list": [ 00:16:13.959 { 00:16:13.959 "name": "BaseBdev1", 00:16:13.959 "uuid": "57f6d7db-955a-4474-b78f-8b9ccf03e227", 00:16:13.959 "is_configured": true, 00:16:13.959 "data_offset": 256, 00:16:13.959 "data_size": 7936 00:16:13.959 }, 00:16:13.959 { 00:16:13.959 "name": "BaseBdev2", 00:16:13.959 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.959 "is_configured": false, 00:16:13.959 "data_offset": 0, 00:16:13.959 "data_size": 0 00:16:13.959 } 00:16:13.959 ] 00:16:13.959 }' 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:13.959 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.220 [2024-09-29 16:54:35.819859] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:14.220 [2024-09-29 16:54:35.819904] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.220 [2024-09-29 16:54:35.831876] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:14.220 [2024-09-29 16:54:35.833667] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:14.220 [2024-09-29 16:54:35.833703] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.220 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:14.220 "name": "Existed_Raid", 00:16:14.220 "uuid": "af52a820-04bc-43c4-9bcd-5cbcbd5662f4", 00:16:14.220 "strip_size_kb": 0, 00:16:14.220 "state": "configuring", 00:16:14.220 "raid_level": "raid1", 00:16:14.220 "superblock": true, 00:16:14.220 "num_base_bdevs": 2, 00:16:14.220 "num_base_bdevs_discovered": 1, 00:16:14.220 "num_base_bdevs_operational": 2, 00:16:14.220 "base_bdevs_list": [ 00:16:14.220 { 00:16:14.220 "name": "BaseBdev1", 00:16:14.220 "uuid": "57f6d7db-955a-4474-b78f-8b9ccf03e227", 00:16:14.220 "is_configured": true, 00:16:14.220 "data_offset": 256, 00:16:14.220 "data_size": 7936 00:16:14.220 }, 00:16:14.221 { 00:16:14.221 "name": "BaseBdev2", 00:16:14.221 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:14.221 "is_configured": false, 00:16:14.221 "data_offset": 0, 00:16:14.221 "data_size": 0 00:16:14.221 } 00:16:14.221 ] 00:16:14.221 }' 00:16:14.221 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:14.221 16:54:35 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.791 [2024-09-29 16:54:36.309500] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:14.791 [2024-09-29 16:54:36.310084] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:14.791 [2024-09-29 16:54:36.310144] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:14.791 [2024-09-29 16:54:36.310504] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:14.791 BaseBdev2 00:16:14.791 [2024-09-29 16:54:36.310893] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:14.791 [2024-09-29 16:54:36.310957] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:14.791 [2024-09-29 16:54:36.311183] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.791 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.791 [ 00:16:14.791 { 00:16:14.791 "name": "BaseBdev2", 00:16:14.791 "aliases": [ 00:16:14.791 "51d84609-4cc8-4a82-957f-050534a46583" 00:16:14.791 ], 00:16:14.791 "product_name": "Malloc disk", 00:16:14.791 "block_size": 4128, 00:16:14.791 "num_blocks": 8192, 00:16:14.791 "uuid": "51d84609-4cc8-4a82-957f-050534a46583", 00:16:14.791 "md_size": 32, 00:16:14.791 "md_interleave": true, 00:16:14.791 "dif_type": 0, 00:16:14.791 "assigned_rate_limits": { 00:16:14.791 "rw_ios_per_sec": 0, 00:16:14.791 "rw_mbytes_per_sec": 0, 00:16:14.791 "r_mbytes_per_sec": 0, 00:16:14.791 "w_mbytes_per_sec": 0 00:16:14.791 }, 00:16:14.791 "claimed": true, 00:16:14.791 "claim_type": "exclusive_write", 00:16:14.791 "zoned": false, 00:16:14.791 "supported_io_types": { 00:16:14.791 "read": true, 00:16:14.791 "write": true, 00:16:14.791 "unmap": true, 00:16:14.791 "flush": true, 00:16:14.791 "reset": true, 00:16:14.791 "nvme_admin": false, 00:16:14.791 "nvme_io": false, 00:16:14.791 "nvme_io_md": false, 00:16:14.791 "write_zeroes": true, 00:16:14.791 "zcopy": true, 00:16:14.791 "get_zone_info": false, 00:16:14.791 "zone_management": false, 00:16:14.791 "zone_append": false, 00:16:14.791 "compare": false, 00:16:14.791 "compare_and_write": false, 00:16:14.791 "abort": true, 00:16:14.791 "seek_hole": false, 00:16:14.791 "seek_data": false, 00:16:14.791 "copy": true, 00:16:14.791 "nvme_iov_md": false 00:16:14.791 }, 00:16:14.791 "memory_domains": [ 00:16:14.791 { 00:16:14.791 "dma_device_id": "system", 00:16:14.791 "dma_device_type": 1 00:16:14.792 }, 00:16:14.792 { 00:16:14.792 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:14.792 "dma_device_type": 2 00:16:14.792 } 00:16:14.792 ], 00:16:14.792 "driver_specific": {} 00:16:14.792 } 00:16:14.792 ] 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:14.792 "name": "Existed_Raid", 00:16:14.792 "uuid": "af52a820-04bc-43c4-9bcd-5cbcbd5662f4", 00:16:14.792 "strip_size_kb": 0, 00:16:14.792 "state": "online", 00:16:14.792 "raid_level": "raid1", 00:16:14.792 "superblock": true, 00:16:14.792 "num_base_bdevs": 2, 00:16:14.792 "num_base_bdevs_discovered": 2, 00:16:14.792 "num_base_bdevs_operational": 2, 00:16:14.792 "base_bdevs_list": [ 00:16:14.792 { 00:16:14.792 "name": "BaseBdev1", 00:16:14.792 "uuid": "57f6d7db-955a-4474-b78f-8b9ccf03e227", 00:16:14.792 "is_configured": true, 00:16:14.792 "data_offset": 256, 00:16:14.792 "data_size": 7936 00:16:14.792 }, 00:16:14.792 { 00:16:14.792 "name": "BaseBdev2", 00:16:14.792 "uuid": "51d84609-4cc8-4a82-957f-050534a46583", 00:16:14.792 "is_configured": true, 00:16:14.792 "data_offset": 256, 00:16:14.792 "data_size": 7936 00:16:14.792 } 00:16:14.792 ] 00:16:14.792 }' 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:14.792 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:15.362 [2024-09-29 16:54:36.773066] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:15.362 "name": "Existed_Raid", 00:16:15.362 "aliases": [ 00:16:15.362 "af52a820-04bc-43c4-9bcd-5cbcbd5662f4" 00:16:15.362 ], 00:16:15.362 "product_name": "Raid Volume", 00:16:15.362 "block_size": 4128, 00:16:15.362 "num_blocks": 7936, 00:16:15.362 "uuid": "af52a820-04bc-43c4-9bcd-5cbcbd5662f4", 00:16:15.362 "md_size": 32, 00:16:15.362 "md_interleave": true, 00:16:15.362 "dif_type": 0, 00:16:15.362 "assigned_rate_limits": { 00:16:15.362 "rw_ios_per_sec": 0, 00:16:15.362 "rw_mbytes_per_sec": 0, 00:16:15.362 "r_mbytes_per_sec": 0, 00:16:15.362 "w_mbytes_per_sec": 0 00:16:15.362 }, 00:16:15.362 "claimed": false, 00:16:15.362 "zoned": false, 00:16:15.362 "supported_io_types": { 00:16:15.362 "read": true, 00:16:15.362 "write": true, 00:16:15.362 "unmap": false, 00:16:15.362 "flush": false, 00:16:15.362 "reset": true, 00:16:15.362 "nvme_admin": false, 00:16:15.362 "nvme_io": false, 00:16:15.362 "nvme_io_md": false, 00:16:15.362 "write_zeroes": true, 00:16:15.362 "zcopy": false, 00:16:15.362 "get_zone_info": false, 00:16:15.362 "zone_management": false, 00:16:15.362 "zone_append": false, 00:16:15.362 "compare": false, 00:16:15.362 "compare_and_write": false, 00:16:15.362 "abort": false, 00:16:15.362 "seek_hole": false, 00:16:15.362 "seek_data": false, 00:16:15.362 "copy": false, 00:16:15.362 "nvme_iov_md": false 00:16:15.362 }, 00:16:15.362 "memory_domains": [ 00:16:15.362 { 00:16:15.362 "dma_device_id": "system", 00:16:15.362 "dma_device_type": 1 00:16:15.362 }, 00:16:15.362 { 00:16:15.362 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:15.362 "dma_device_type": 2 00:16:15.362 }, 00:16:15.362 { 00:16:15.362 "dma_device_id": "system", 00:16:15.362 "dma_device_type": 1 00:16:15.362 }, 00:16:15.362 { 00:16:15.362 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:15.362 "dma_device_type": 2 00:16:15.362 } 00:16:15.362 ], 00:16:15.362 "driver_specific": { 00:16:15.362 "raid": { 00:16:15.362 "uuid": "af52a820-04bc-43c4-9bcd-5cbcbd5662f4", 00:16:15.362 "strip_size_kb": 0, 00:16:15.362 "state": "online", 00:16:15.362 "raid_level": "raid1", 00:16:15.362 "superblock": true, 00:16:15.362 "num_base_bdevs": 2, 00:16:15.362 "num_base_bdevs_discovered": 2, 00:16:15.362 "num_base_bdevs_operational": 2, 00:16:15.362 "base_bdevs_list": [ 00:16:15.362 { 00:16:15.362 "name": "BaseBdev1", 00:16:15.362 "uuid": "57f6d7db-955a-4474-b78f-8b9ccf03e227", 00:16:15.362 "is_configured": true, 00:16:15.362 "data_offset": 256, 00:16:15.362 "data_size": 7936 00:16:15.362 }, 00:16:15.362 { 00:16:15.362 "name": "BaseBdev2", 00:16:15.362 "uuid": "51d84609-4cc8-4a82-957f-050534a46583", 00:16:15.362 "is_configured": true, 00:16:15.362 "data_offset": 256, 00:16:15.362 "data_size": 7936 00:16:15.362 } 00:16:15.362 ] 00:16:15.362 } 00:16:15.362 } 00:16:15.362 }' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:15.362 BaseBdev2' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:15.362 16:54:36 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.362 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:15.362 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:15.362 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.363 [2024-09-29 16:54:37.008480] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:15.363 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.623 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.623 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:15.623 "name": "Existed_Raid", 00:16:15.623 "uuid": "af52a820-04bc-43c4-9bcd-5cbcbd5662f4", 00:16:15.623 "strip_size_kb": 0, 00:16:15.623 "state": "online", 00:16:15.623 "raid_level": "raid1", 00:16:15.623 "superblock": true, 00:16:15.623 "num_base_bdevs": 2, 00:16:15.623 "num_base_bdevs_discovered": 1, 00:16:15.623 "num_base_bdevs_operational": 1, 00:16:15.623 "base_bdevs_list": [ 00:16:15.623 { 00:16:15.623 "name": null, 00:16:15.623 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:15.623 "is_configured": false, 00:16:15.623 "data_offset": 0, 00:16:15.623 "data_size": 7936 00:16:15.623 }, 00:16:15.623 { 00:16:15.623 "name": "BaseBdev2", 00:16:15.623 "uuid": "51d84609-4cc8-4a82-957f-050534a46583", 00:16:15.623 "is_configured": true, 00:16:15.623 "data_offset": 256, 00:16:15.623 "data_size": 7936 00:16:15.623 } 00:16:15.623 ] 00:16:15.623 }' 00:16:15.623 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:15.623 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.883 [2024-09-29 16:54:37.511469] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:15.883 [2024-09-29 16:54:37.511570] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:15.883 [2024-09-29 16:54:37.523558] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:15.883 [2024-09-29 16:54:37.523605] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:15.883 [2024-09-29 16:54:37.523616] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:15.883 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 98407 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 98407 ']' 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 98407 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 98407 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:16.143 killing process with pid 98407 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 98407' 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 98407 00:16:16.143 [2024-09-29 16:54:37.618219] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:16.143 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 98407 00:16:16.143 [2024-09-29 16:54:37.619222] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:16.405 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:16:16.405 00:16:16.405 real 0m3.856s 00:16:16.405 user 0m6.027s 00:16:16.405 sys 0m0.835s 00:16:16.405 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:16.405 16:54:37 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.405 ************************************ 00:16:16.405 END TEST raid_state_function_test_sb_md_interleaved 00:16:16.405 ************************************ 00:16:16.405 16:54:37 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:16:16.405 16:54:37 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:16:16.405 16:54:37 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:16.405 16:54:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:16.405 ************************************ 00:16:16.405 START TEST raid_superblock_test_md_interleaved 00:16:16.405 ************************************ 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=98649 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 98649 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 98649 ']' 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:16.405 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:16.405 16:54:37 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:16.405 [2024-09-29 16:54:38.035477] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:16.405 [2024-09-29 16:54:38.035597] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98649 ] 00:16:16.665 [2024-09-29 16:54:38.180464] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:16.665 [2024-09-29 16:54:38.227293] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:16.665 [2024-09-29 16:54:38.270417] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:16.665 [2024-09-29 16:54:38.270462] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.236 malloc1 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.236 [2024-09-29 16:54:38.861214] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:17.236 [2024-09-29 16:54:38.861277] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:17.236 [2024-09-29 16:54:38.861299] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:17.236 [2024-09-29 16:54:38.861327] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:17.236 [2024-09-29 16:54:38.863251] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:17.236 [2024-09-29 16:54:38.863298] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:17.236 pt1 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.236 malloc2 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.236 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.501 [2024-09-29 16:54:38.909272] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:17.501 [2024-09-29 16:54:38.909371] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:17.501 [2024-09-29 16:54:38.909416] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:17.501 [2024-09-29 16:54:38.909443] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:17.501 [2024-09-29 16:54:38.913806] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:17.501 [2024-09-29 16:54:38.913876] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:17.501 pt2 00:16:17.501 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.501 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:17.501 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:17.501 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:17.501 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.501 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.501 [2024-09-29 16:54:38.922169] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:17.502 [2024-09-29 16:54:38.925113] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:17.502 [2024-09-29 16:54:38.925379] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:17.502 [2024-09-29 16:54:38.925422] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:17.502 [2024-09-29 16:54:38.925579] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:17.502 [2024-09-29 16:54:38.925779] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:17.502 [2024-09-29 16:54:38.925806] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:17.502 [2024-09-29 16:54:38.925931] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:17.502 "name": "raid_bdev1", 00:16:17.502 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:17.502 "strip_size_kb": 0, 00:16:17.502 "state": "online", 00:16:17.502 "raid_level": "raid1", 00:16:17.502 "superblock": true, 00:16:17.502 "num_base_bdevs": 2, 00:16:17.502 "num_base_bdevs_discovered": 2, 00:16:17.502 "num_base_bdevs_operational": 2, 00:16:17.502 "base_bdevs_list": [ 00:16:17.502 { 00:16:17.502 "name": "pt1", 00:16:17.502 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:17.502 "is_configured": true, 00:16:17.502 "data_offset": 256, 00:16:17.502 "data_size": 7936 00:16:17.502 }, 00:16:17.502 { 00:16:17.502 "name": "pt2", 00:16:17.502 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:17.502 "is_configured": true, 00:16:17.502 "data_offset": 256, 00:16:17.502 "data_size": 7936 00:16:17.502 } 00:16:17.502 ] 00:16:17.502 }' 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:17.502 16:54:38 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:17.763 [2024-09-29 16:54:39.369514] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.763 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:17.763 "name": "raid_bdev1", 00:16:17.763 "aliases": [ 00:16:17.763 "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753" 00:16:17.763 ], 00:16:17.763 "product_name": "Raid Volume", 00:16:17.763 "block_size": 4128, 00:16:17.763 "num_blocks": 7936, 00:16:17.763 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:17.763 "md_size": 32, 00:16:17.763 "md_interleave": true, 00:16:17.763 "dif_type": 0, 00:16:17.763 "assigned_rate_limits": { 00:16:17.763 "rw_ios_per_sec": 0, 00:16:17.763 "rw_mbytes_per_sec": 0, 00:16:17.763 "r_mbytes_per_sec": 0, 00:16:17.763 "w_mbytes_per_sec": 0 00:16:17.763 }, 00:16:17.763 "claimed": false, 00:16:17.763 "zoned": false, 00:16:17.763 "supported_io_types": { 00:16:17.763 "read": true, 00:16:17.763 "write": true, 00:16:17.763 "unmap": false, 00:16:17.763 "flush": false, 00:16:17.763 "reset": true, 00:16:17.763 "nvme_admin": false, 00:16:17.763 "nvme_io": false, 00:16:17.763 "nvme_io_md": false, 00:16:17.763 "write_zeroes": true, 00:16:17.763 "zcopy": false, 00:16:17.763 "get_zone_info": false, 00:16:17.763 "zone_management": false, 00:16:17.763 "zone_append": false, 00:16:17.763 "compare": false, 00:16:17.763 "compare_and_write": false, 00:16:17.763 "abort": false, 00:16:17.763 "seek_hole": false, 00:16:17.763 "seek_data": false, 00:16:17.763 "copy": false, 00:16:17.763 "nvme_iov_md": false 00:16:17.763 }, 00:16:17.763 "memory_domains": [ 00:16:17.763 { 00:16:17.763 "dma_device_id": "system", 00:16:17.763 "dma_device_type": 1 00:16:17.763 }, 00:16:17.763 { 00:16:17.763 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:17.763 "dma_device_type": 2 00:16:17.763 }, 00:16:17.763 { 00:16:17.763 "dma_device_id": "system", 00:16:17.763 "dma_device_type": 1 00:16:17.763 }, 00:16:17.763 { 00:16:17.763 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:17.763 "dma_device_type": 2 00:16:17.763 } 00:16:17.763 ], 00:16:17.763 "driver_specific": { 00:16:17.763 "raid": { 00:16:17.763 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:17.763 "strip_size_kb": 0, 00:16:17.763 "state": "online", 00:16:17.763 "raid_level": "raid1", 00:16:17.763 "superblock": true, 00:16:17.763 "num_base_bdevs": 2, 00:16:17.763 "num_base_bdevs_discovered": 2, 00:16:17.763 "num_base_bdevs_operational": 2, 00:16:17.763 "base_bdevs_list": [ 00:16:17.763 { 00:16:17.763 "name": "pt1", 00:16:17.763 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:17.763 "is_configured": true, 00:16:17.763 "data_offset": 256, 00:16:17.763 "data_size": 7936 00:16:17.763 }, 00:16:17.763 { 00:16:17.763 "name": "pt2", 00:16:17.764 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:17.764 "is_configured": true, 00:16:17.764 "data_offset": 256, 00:16:17.764 "data_size": 7936 00:16:17.764 } 00:16:17.764 ] 00:16:17.764 } 00:16:17.764 } 00:16:17.764 }' 00:16:17.764 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:18.022 pt2' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.022 [2024-09-29 16:54:39.613029] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=fc9e62a9-42c0-47ad-8b30-8aaaa35d0753 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z fc9e62a9-42c0-47ad-8b30-8aaaa35d0753 ']' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.022 [2024-09-29 16:54:39.640791] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:18.022 [2024-09-29 16:54:39.640817] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:18.022 [2024-09-29 16:54:39.640878] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:18.022 [2024-09-29 16:54:39.640937] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:18.022 [2024-09-29 16:54:39.640946] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.022 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.282 [2024-09-29 16:54:39.776572] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:18.282 [2024-09-29 16:54:39.778365] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:18.282 [2024-09-29 16:54:39.778425] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:18.282 [2024-09-29 16:54:39.778471] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:18.282 [2024-09-29 16:54:39.778487] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:18.282 [2024-09-29 16:54:39.778496] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:18.282 request: 00:16:18.282 { 00:16:18.282 "name": "raid_bdev1", 00:16:18.282 "raid_level": "raid1", 00:16:18.282 "base_bdevs": [ 00:16:18.282 "malloc1", 00:16:18.282 "malloc2" 00:16:18.282 ], 00:16:18.282 "superblock": false, 00:16:18.282 "method": "bdev_raid_create", 00:16:18.282 "req_id": 1 00:16:18.282 } 00:16:18.282 Got JSON-RPC error response 00:16:18.282 response: 00:16:18.282 { 00:16:18.282 "code": -17, 00:16:18.282 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:18.282 } 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.282 [2024-09-29 16:54:39.828450] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:18.282 [2024-09-29 16:54:39.828492] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:18.282 [2024-09-29 16:54:39.828508] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:18.282 [2024-09-29 16:54:39.828516] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:18.282 [2024-09-29 16:54:39.830345] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:18.282 [2024-09-29 16:54:39.830375] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:18.282 [2024-09-29 16:54:39.830414] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:18.282 [2024-09-29 16:54:39.830442] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:18.282 pt1 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:18.282 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.283 "name": "raid_bdev1", 00:16:18.283 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:18.283 "strip_size_kb": 0, 00:16:18.283 "state": "configuring", 00:16:18.283 "raid_level": "raid1", 00:16:18.283 "superblock": true, 00:16:18.283 "num_base_bdevs": 2, 00:16:18.283 "num_base_bdevs_discovered": 1, 00:16:18.283 "num_base_bdevs_operational": 2, 00:16:18.283 "base_bdevs_list": [ 00:16:18.283 { 00:16:18.283 "name": "pt1", 00:16:18.283 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:18.283 "is_configured": true, 00:16:18.283 "data_offset": 256, 00:16:18.283 "data_size": 7936 00:16:18.283 }, 00:16:18.283 { 00:16:18.283 "name": null, 00:16:18.283 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:18.283 "is_configured": false, 00:16:18.283 "data_offset": 256, 00:16:18.283 "data_size": 7936 00:16:18.283 } 00:16:18.283 ] 00:16:18.283 }' 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.283 16:54:39 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.852 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:18.852 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:18.852 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:18.852 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:18.852 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.852 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.852 [2024-09-29 16:54:40.299678] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:18.852 [2024-09-29 16:54:40.299731] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:18.853 [2024-09-29 16:54:40.299750] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:18.853 [2024-09-29 16:54:40.299758] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:18.853 [2024-09-29 16:54:40.299877] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:18.853 [2024-09-29 16:54:40.299891] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:18.853 [2024-09-29 16:54:40.299928] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:18.853 [2024-09-29 16:54:40.299949] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:18.853 [2024-09-29 16:54:40.300018] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:18.853 [2024-09-29 16:54:40.300026] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:18.853 [2024-09-29 16:54:40.300088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:18.853 [2024-09-29 16:54:40.300136] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:18.853 [2024-09-29 16:54:40.300148] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:16:18.853 [2024-09-29 16:54:40.300194] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:18.853 pt2 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.853 "name": "raid_bdev1", 00:16:18.853 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:18.853 "strip_size_kb": 0, 00:16:18.853 "state": "online", 00:16:18.853 "raid_level": "raid1", 00:16:18.853 "superblock": true, 00:16:18.853 "num_base_bdevs": 2, 00:16:18.853 "num_base_bdevs_discovered": 2, 00:16:18.853 "num_base_bdevs_operational": 2, 00:16:18.853 "base_bdevs_list": [ 00:16:18.853 { 00:16:18.853 "name": "pt1", 00:16:18.853 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:18.853 "is_configured": true, 00:16:18.853 "data_offset": 256, 00:16:18.853 "data_size": 7936 00:16:18.853 }, 00:16:18.853 { 00:16:18.853 "name": "pt2", 00:16:18.853 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:18.853 "is_configured": true, 00:16:18.853 "data_offset": 256, 00:16:18.853 "data_size": 7936 00:16:18.853 } 00:16:18.853 ] 00:16:18.853 }' 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.853 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.113 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.113 [2024-09-29 16:54:40.779224] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:19.374 "name": "raid_bdev1", 00:16:19.374 "aliases": [ 00:16:19.374 "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753" 00:16:19.374 ], 00:16:19.374 "product_name": "Raid Volume", 00:16:19.374 "block_size": 4128, 00:16:19.374 "num_blocks": 7936, 00:16:19.374 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:19.374 "md_size": 32, 00:16:19.374 "md_interleave": true, 00:16:19.374 "dif_type": 0, 00:16:19.374 "assigned_rate_limits": { 00:16:19.374 "rw_ios_per_sec": 0, 00:16:19.374 "rw_mbytes_per_sec": 0, 00:16:19.374 "r_mbytes_per_sec": 0, 00:16:19.374 "w_mbytes_per_sec": 0 00:16:19.374 }, 00:16:19.374 "claimed": false, 00:16:19.374 "zoned": false, 00:16:19.374 "supported_io_types": { 00:16:19.374 "read": true, 00:16:19.374 "write": true, 00:16:19.374 "unmap": false, 00:16:19.374 "flush": false, 00:16:19.374 "reset": true, 00:16:19.374 "nvme_admin": false, 00:16:19.374 "nvme_io": false, 00:16:19.374 "nvme_io_md": false, 00:16:19.374 "write_zeroes": true, 00:16:19.374 "zcopy": false, 00:16:19.374 "get_zone_info": false, 00:16:19.374 "zone_management": false, 00:16:19.374 "zone_append": false, 00:16:19.374 "compare": false, 00:16:19.374 "compare_and_write": false, 00:16:19.374 "abort": false, 00:16:19.374 "seek_hole": false, 00:16:19.374 "seek_data": false, 00:16:19.374 "copy": false, 00:16:19.374 "nvme_iov_md": false 00:16:19.374 }, 00:16:19.374 "memory_domains": [ 00:16:19.374 { 00:16:19.374 "dma_device_id": "system", 00:16:19.374 "dma_device_type": 1 00:16:19.374 }, 00:16:19.374 { 00:16:19.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:19.374 "dma_device_type": 2 00:16:19.374 }, 00:16:19.374 { 00:16:19.374 "dma_device_id": "system", 00:16:19.374 "dma_device_type": 1 00:16:19.374 }, 00:16:19.374 { 00:16:19.374 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:19.374 "dma_device_type": 2 00:16:19.374 } 00:16:19.374 ], 00:16:19.374 "driver_specific": { 00:16:19.374 "raid": { 00:16:19.374 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:19.374 "strip_size_kb": 0, 00:16:19.374 "state": "online", 00:16:19.374 "raid_level": "raid1", 00:16:19.374 "superblock": true, 00:16:19.374 "num_base_bdevs": 2, 00:16:19.374 "num_base_bdevs_discovered": 2, 00:16:19.374 "num_base_bdevs_operational": 2, 00:16:19.374 "base_bdevs_list": [ 00:16:19.374 { 00:16:19.374 "name": "pt1", 00:16:19.374 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:19.374 "is_configured": true, 00:16:19.374 "data_offset": 256, 00:16:19.374 "data_size": 7936 00:16:19.374 }, 00:16:19.374 { 00:16:19.374 "name": "pt2", 00:16:19.374 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:19.374 "is_configured": true, 00:16:19.374 "data_offset": 256, 00:16:19.374 "data_size": 7936 00:16:19.374 } 00:16:19.374 ] 00:16:19.374 } 00:16:19.374 } 00:16:19.374 }' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:19.374 pt2' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:19.374 [2024-09-29 16:54:40.978935] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:19.374 16:54:40 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' fc9e62a9-42c0-47ad-8b30-8aaaa35d0753 '!=' fc9e62a9-42c0-47ad-8b30-8aaaa35d0753 ']' 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.374 [2024-09-29 16:54:41.022653] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.374 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.634 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.634 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:19.634 "name": "raid_bdev1", 00:16:19.634 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:19.634 "strip_size_kb": 0, 00:16:19.634 "state": "online", 00:16:19.634 "raid_level": "raid1", 00:16:19.634 "superblock": true, 00:16:19.634 "num_base_bdevs": 2, 00:16:19.634 "num_base_bdevs_discovered": 1, 00:16:19.634 "num_base_bdevs_operational": 1, 00:16:19.634 "base_bdevs_list": [ 00:16:19.634 { 00:16:19.634 "name": null, 00:16:19.634 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:19.634 "is_configured": false, 00:16:19.634 "data_offset": 0, 00:16:19.634 "data_size": 7936 00:16:19.634 }, 00:16:19.634 { 00:16:19.634 "name": "pt2", 00:16:19.634 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:19.634 "is_configured": true, 00:16:19.634 "data_offset": 256, 00:16:19.634 "data_size": 7936 00:16:19.634 } 00:16:19.634 ] 00:16:19.634 }' 00:16:19.634 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:19.634 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.894 [2024-09-29 16:54:41.453873] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:19.894 [2024-09-29 16:54:41.453902] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:19.894 [2024-09-29 16:54:41.453952] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:19.894 [2024-09-29 16:54:41.453992] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:19.894 [2024-09-29 16:54:41.454001] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.894 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.895 [2024-09-29 16:54:41.525761] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:19.895 [2024-09-29 16:54:41.525804] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:19.895 [2024-09-29 16:54:41.525823] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:16:19.895 [2024-09-29 16:54:41.525831] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:19.895 [2024-09-29 16:54:41.527660] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:19.895 [2024-09-29 16:54:41.527688] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:19.895 [2024-09-29 16:54:41.527741] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:19.895 [2024-09-29 16:54:41.527767] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:19.895 [2024-09-29 16:54:41.527817] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:19.895 [2024-09-29 16:54:41.527825] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:19.895 [2024-09-29 16:54:41.527889] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:19.895 [2024-09-29 16:54:41.527940] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:19.895 [2024-09-29 16:54:41.527948] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:19.895 [2024-09-29 16:54:41.527995] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:19.895 pt2 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:19.895 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.155 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.155 "name": "raid_bdev1", 00:16:20.155 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:20.155 "strip_size_kb": 0, 00:16:20.155 "state": "online", 00:16:20.155 "raid_level": "raid1", 00:16:20.155 "superblock": true, 00:16:20.155 "num_base_bdevs": 2, 00:16:20.155 "num_base_bdevs_discovered": 1, 00:16:20.155 "num_base_bdevs_operational": 1, 00:16:20.155 "base_bdevs_list": [ 00:16:20.155 { 00:16:20.155 "name": null, 00:16:20.155 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.155 "is_configured": false, 00:16:20.155 "data_offset": 256, 00:16:20.155 "data_size": 7936 00:16:20.155 }, 00:16:20.155 { 00:16:20.155 "name": "pt2", 00:16:20.155 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:20.155 "is_configured": true, 00:16:20.155 "data_offset": 256, 00:16:20.155 "data_size": 7936 00:16:20.155 } 00:16:20.155 ] 00:16:20.155 }' 00:16:20.155 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.155 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.415 [2024-09-29 16:54:41.988941] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:20.415 [2024-09-29 16:54:41.988966] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:20.415 [2024-09-29 16:54:41.989022] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:20.415 [2024-09-29 16:54:41.989059] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:20.415 [2024-09-29 16:54:41.989070] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.415 16:54:41 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.415 [2024-09-29 16:54:42.048856] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:20.415 [2024-09-29 16:54:42.048900] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:20.415 [2024-09-29 16:54:42.048915] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:16:20.415 [2024-09-29 16:54:42.048926] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:20.415 [2024-09-29 16:54:42.050771] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:20.415 [2024-09-29 16:54:42.050801] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:20.415 [2024-09-29 16:54:42.050839] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:20.415 [2024-09-29 16:54:42.050867] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:20.415 [2024-09-29 16:54:42.050963] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:20.415 [2024-09-29 16:54:42.050977] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:20.415 [2024-09-29 16:54:42.050991] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:20.415 [2024-09-29 16:54:42.051016] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:20.415 [2024-09-29 16:54:42.051063] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:20.415 [2024-09-29 16:54:42.051072] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:20.415 [2024-09-29 16:54:42.051157] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:20.415 [2024-09-29 16:54:42.051226] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:20.415 [2024-09-29 16:54:42.051233] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:20.415 [2024-09-29 16:54:42.051292] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:20.415 pt1 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.415 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.675 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.675 "name": "raid_bdev1", 00:16:20.675 "uuid": "fc9e62a9-42c0-47ad-8b30-8aaaa35d0753", 00:16:20.675 "strip_size_kb": 0, 00:16:20.675 "state": "online", 00:16:20.675 "raid_level": "raid1", 00:16:20.675 "superblock": true, 00:16:20.675 "num_base_bdevs": 2, 00:16:20.675 "num_base_bdevs_discovered": 1, 00:16:20.675 "num_base_bdevs_operational": 1, 00:16:20.675 "base_bdevs_list": [ 00:16:20.675 { 00:16:20.675 "name": null, 00:16:20.675 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.675 "is_configured": false, 00:16:20.675 "data_offset": 256, 00:16:20.675 "data_size": 7936 00:16:20.675 }, 00:16:20.675 { 00:16:20.675 "name": "pt2", 00:16:20.675 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:20.675 "is_configured": true, 00:16:20.675 "data_offset": 256, 00:16:20.675 "data_size": 7936 00:16:20.675 } 00:16:20.675 ] 00:16:20.675 }' 00:16:20.675 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.675 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:20.935 [2024-09-29 16:54:42.552201] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' fc9e62a9-42c0-47ad-8b30-8aaaa35d0753 '!=' fc9e62a9-42c0-47ad-8b30-8aaaa35d0753 ']' 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 98649 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 98649 ']' 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 98649 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:20.935 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 98649 00:16:21.195 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:21.195 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:21.195 killing process with pid 98649 00:16:21.195 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 98649' 00:16:21.195 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@969 -- # kill 98649 00:16:21.195 [2024-09-29 16:54:42.626994] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:21.195 [2024-09-29 16:54:42.627053] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:21.195 [2024-09-29 16:54:42.627103] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:21.195 [2024-09-29 16:54:42.627115] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:21.195 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@974 -- # wait 98649 00:16:21.195 [2024-09-29 16:54:42.650914] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:21.455 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:16:21.455 00:16:21.455 real 0m4.946s 00:16:21.455 user 0m8.048s 00:16:21.455 sys 0m1.118s 00:16:21.455 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:21.455 16:54:42 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.455 ************************************ 00:16:21.455 END TEST raid_superblock_test_md_interleaved 00:16:21.455 ************************************ 00:16:21.455 16:54:42 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:16:21.455 16:54:42 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:16:21.455 16:54:42 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:21.455 16:54:42 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:21.455 ************************************ 00:16:21.455 START TEST raid_rebuild_test_sb_md_interleaved 00:16:21.455 ************************************ 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false false 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=98961 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 98961 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 98961 ']' 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:21.455 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:21.455 16:54:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:21.455 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:21.455 Zero copy mechanism will not be used. 00:16:21.455 [2024-09-29 16:54:43.075605] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:21.455 [2024-09-29 16:54:43.075747] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98961 ] 00:16:21.715 [2024-09-29 16:54:43.219880] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:21.716 [2024-09-29 16:54:43.266703] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:21.716 [2024-09-29 16:54:43.310186] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:21.716 [2024-09-29 16:54:43.310222] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.286 BaseBdev1_malloc 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.286 [2024-09-29 16:54:43.900703] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:22.286 [2024-09-29 16:54:43.900792] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:22.286 [2024-09-29 16:54:43.900816] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:22.286 [2024-09-29 16:54:43.900825] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:22.286 [2024-09-29 16:54:43.902597] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:22.286 [2024-09-29 16:54:43.902639] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:22.286 BaseBdev1 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.286 BaseBdev2_malloc 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.286 [2024-09-29 16:54:43.938360] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:22.286 [2024-09-29 16:54:43.938443] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:22.286 [2024-09-29 16:54:43.938483] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:22.286 [2024-09-29 16:54:43.938504] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:22.286 [2024-09-29 16:54:43.942545] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:22.286 [2024-09-29 16:54:43.942603] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:22.286 BaseBdev2 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.286 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.546 spare_malloc 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.546 spare_delay 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.546 [2024-09-29 16:54:43.981104] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:22.546 [2024-09-29 16:54:43.981156] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:22.546 [2024-09-29 16:54:43.981177] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:22.546 [2024-09-29 16:54:43.981185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:22.546 [2024-09-29 16:54:43.983057] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:22.546 [2024-09-29 16:54:43.983087] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:22.546 spare 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.546 [2024-09-29 16:54:43.993132] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:22.546 [2024-09-29 16:54:43.994912] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:22.546 [2024-09-29 16:54:43.995064] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:22.546 [2024-09-29 16:54:43.995077] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:22.546 [2024-09-29 16:54:43.995161] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:22.546 [2024-09-29 16:54:43.995245] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:22.546 [2024-09-29 16:54:43.995266] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:22.546 [2024-09-29 16:54:43.995337] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:22.546 16:54:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:22.546 "name": "raid_bdev1", 00:16:22.546 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:22.546 "strip_size_kb": 0, 00:16:22.546 "state": "online", 00:16:22.546 "raid_level": "raid1", 00:16:22.546 "superblock": true, 00:16:22.546 "num_base_bdevs": 2, 00:16:22.546 "num_base_bdevs_discovered": 2, 00:16:22.546 "num_base_bdevs_operational": 2, 00:16:22.546 "base_bdevs_list": [ 00:16:22.546 { 00:16:22.546 "name": "BaseBdev1", 00:16:22.546 "uuid": "c99d1f59-041a-5a64-a471-05e730f8bff2", 00:16:22.546 "is_configured": true, 00:16:22.546 "data_offset": 256, 00:16:22.546 "data_size": 7936 00:16:22.546 }, 00:16:22.546 { 00:16:22.546 "name": "BaseBdev2", 00:16:22.546 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:22.546 "is_configured": true, 00:16:22.546 "data_offset": 256, 00:16:22.546 "data_size": 7936 00:16:22.546 } 00:16:22.546 ] 00:16:22.546 }' 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:22.546 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.806 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:22.806 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:22.806 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.806 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:22.806 [2024-09-29 16:54:44.472649] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.066 [2024-09-29 16:54:44.552258] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:23.066 "name": "raid_bdev1", 00:16:23.066 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:23.066 "strip_size_kb": 0, 00:16:23.066 "state": "online", 00:16:23.066 "raid_level": "raid1", 00:16:23.066 "superblock": true, 00:16:23.066 "num_base_bdevs": 2, 00:16:23.066 "num_base_bdevs_discovered": 1, 00:16:23.066 "num_base_bdevs_operational": 1, 00:16:23.066 "base_bdevs_list": [ 00:16:23.066 { 00:16:23.066 "name": null, 00:16:23.066 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:23.066 "is_configured": false, 00:16:23.066 "data_offset": 0, 00:16:23.066 "data_size": 7936 00:16:23.066 }, 00:16:23.066 { 00:16:23.066 "name": "BaseBdev2", 00:16:23.066 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:23.066 "is_configured": true, 00:16:23.066 "data_offset": 256, 00:16:23.066 "data_size": 7936 00:16:23.066 } 00:16:23.066 ] 00:16:23.066 }' 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:23.066 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.326 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:23.326 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.326 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:23.326 [2024-09-29 16:54:44.987514] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:23.326 [2024-09-29 16:54:44.990453] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:23.326 [2024-09-29 16:54:44.992341] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:23.326 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.326 16:54:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:24.707 16:54:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:24.707 16:54:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:24.707 16:54:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:24.707 16:54:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:24.707 16:54:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:24.708 "name": "raid_bdev1", 00:16:24.708 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:24.708 "strip_size_kb": 0, 00:16:24.708 "state": "online", 00:16:24.708 "raid_level": "raid1", 00:16:24.708 "superblock": true, 00:16:24.708 "num_base_bdevs": 2, 00:16:24.708 "num_base_bdevs_discovered": 2, 00:16:24.708 "num_base_bdevs_operational": 2, 00:16:24.708 "process": { 00:16:24.708 "type": "rebuild", 00:16:24.708 "target": "spare", 00:16:24.708 "progress": { 00:16:24.708 "blocks": 2560, 00:16:24.708 "percent": 32 00:16:24.708 } 00:16:24.708 }, 00:16:24.708 "base_bdevs_list": [ 00:16:24.708 { 00:16:24.708 "name": "spare", 00:16:24.708 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:24.708 "is_configured": true, 00:16:24.708 "data_offset": 256, 00:16:24.708 "data_size": 7936 00:16:24.708 }, 00:16:24.708 { 00:16:24.708 "name": "BaseBdev2", 00:16:24.708 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:24.708 "is_configured": true, 00:16:24.708 "data_offset": 256, 00:16:24.708 "data_size": 7936 00:16:24.708 } 00:16:24.708 ] 00:16:24.708 }' 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.708 [2024-09-29 16:54:46.155081] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:24.708 [2024-09-29 16:54:46.197050] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:24.708 [2024-09-29 16:54:46.197105] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:24.708 [2024-09-29 16:54:46.197120] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:24.708 [2024-09-29 16:54:46.197127] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:24.708 "name": "raid_bdev1", 00:16:24.708 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:24.708 "strip_size_kb": 0, 00:16:24.708 "state": "online", 00:16:24.708 "raid_level": "raid1", 00:16:24.708 "superblock": true, 00:16:24.708 "num_base_bdevs": 2, 00:16:24.708 "num_base_bdevs_discovered": 1, 00:16:24.708 "num_base_bdevs_operational": 1, 00:16:24.708 "base_bdevs_list": [ 00:16:24.708 { 00:16:24.708 "name": null, 00:16:24.708 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.708 "is_configured": false, 00:16:24.708 "data_offset": 0, 00:16:24.708 "data_size": 7936 00:16:24.708 }, 00:16:24.708 { 00:16:24.708 "name": "BaseBdev2", 00:16:24.708 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:24.708 "is_configured": true, 00:16:24.708 "data_offset": 256, 00:16:24.708 "data_size": 7936 00:16:24.708 } 00:16:24.708 ] 00:16:24.708 }' 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:24.708 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:25.278 "name": "raid_bdev1", 00:16:25.278 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:25.278 "strip_size_kb": 0, 00:16:25.278 "state": "online", 00:16:25.278 "raid_level": "raid1", 00:16:25.278 "superblock": true, 00:16:25.278 "num_base_bdevs": 2, 00:16:25.278 "num_base_bdevs_discovered": 1, 00:16:25.278 "num_base_bdevs_operational": 1, 00:16:25.278 "base_bdevs_list": [ 00:16:25.278 { 00:16:25.278 "name": null, 00:16:25.278 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:25.278 "is_configured": false, 00:16:25.278 "data_offset": 0, 00:16:25.278 "data_size": 7936 00:16:25.278 }, 00:16:25.278 { 00:16:25.278 "name": "BaseBdev2", 00:16:25.278 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:25.278 "is_configured": true, 00:16:25.278 "data_offset": 256, 00:16:25.278 "data_size": 7936 00:16:25.278 } 00:16:25.278 ] 00:16:25.278 }' 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:25.278 [2024-09-29 16:54:46.771509] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:25.278 [2024-09-29 16:54:46.774129] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:25.278 [2024-09-29 16:54:46.775990] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.278 16:54:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.218 "name": "raid_bdev1", 00:16:26.218 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:26.218 "strip_size_kb": 0, 00:16:26.218 "state": "online", 00:16:26.218 "raid_level": "raid1", 00:16:26.218 "superblock": true, 00:16:26.218 "num_base_bdevs": 2, 00:16:26.218 "num_base_bdevs_discovered": 2, 00:16:26.218 "num_base_bdevs_operational": 2, 00:16:26.218 "process": { 00:16:26.218 "type": "rebuild", 00:16:26.218 "target": "spare", 00:16:26.218 "progress": { 00:16:26.218 "blocks": 2560, 00:16:26.218 "percent": 32 00:16:26.218 } 00:16:26.218 }, 00:16:26.218 "base_bdevs_list": [ 00:16:26.218 { 00:16:26.218 "name": "spare", 00:16:26.218 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:26.218 "is_configured": true, 00:16:26.218 "data_offset": 256, 00:16:26.218 "data_size": 7936 00:16:26.218 }, 00:16:26.218 { 00:16:26.218 "name": "BaseBdev2", 00:16:26.218 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:26.218 "is_configured": true, 00:16:26.218 "data_offset": 256, 00:16:26.218 "data_size": 7936 00:16:26.218 } 00:16:26.218 ] 00:16:26.218 }' 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:26.218 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:26.478 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=612 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.478 "name": "raid_bdev1", 00:16:26.478 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:26.478 "strip_size_kb": 0, 00:16:26.478 "state": "online", 00:16:26.478 "raid_level": "raid1", 00:16:26.478 "superblock": true, 00:16:26.478 "num_base_bdevs": 2, 00:16:26.478 "num_base_bdevs_discovered": 2, 00:16:26.478 "num_base_bdevs_operational": 2, 00:16:26.478 "process": { 00:16:26.478 "type": "rebuild", 00:16:26.478 "target": "spare", 00:16:26.478 "progress": { 00:16:26.478 "blocks": 2816, 00:16:26.478 "percent": 35 00:16:26.478 } 00:16:26.478 }, 00:16:26.478 "base_bdevs_list": [ 00:16:26.478 { 00:16:26.478 "name": "spare", 00:16:26.478 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:26.478 "is_configured": true, 00:16:26.478 "data_offset": 256, 00:16:26.478 "data_size": 7936 00:16:26.478 }, 00:16:26.478 { 00:16:26.478 "name": "BaseBdev2", 00:16:26.478 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:26.478 "is_configured": true, 00:16:26.478 "data_offset": 256, 00:16:26.478 "data_size": 7936 00:16:26.478 } 00:16:26.478 ] 00:16:26.478 }' 00:16:26.478 16:54:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.478 16:54:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:26.478 16:54:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.479 16:54:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:26.479 16:54:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:27.417 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.675 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:27.675 "name": "raid_bdev1", 00:16:27.675 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:27.675 "strip_size_kb": 0, 00:16:27.675 "state": "online", 00:16:27.675 "raid_level": "raid1", 00:16:27.675 "superblock": true, 00:16:27.675 "num_base_bdevs": 2, 00:16:27.675 "num_base_bdevs_discovered": 2, 00:16:27.675 "num_base_bdevs_operational": 2, 00:16:27.675 "process": { 00:16:27.675 "type": "rebuild", 00:16:27.675 "target": "spare", 00:16:27.675 "progress": { 00:16:27.675 "blocks": 5632, 00:16:27.675 "percent": 70 00:16:27.675 } 00:16:27.675 }, 00:16:27.675 "base_bdevs_list": [ 00:16:27.675 { 00:16:27.675 "name": "spare", 00:16:27.675 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:27.675 "is_configured": true, 00:16:27.675 "data_offset": 256, 00:16:27.675 "data_size": 7936 00:16:27.675 }, 00:16:27.675 { 00:16:27.675 "name": "BaseBdev2", 00:16:27.675 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:27.675 "is_configured": true, 00:16:27.675 "data_offset": 256, 00:16:27.675 "data_size": 7936 00:16:27.675 } 00:16:27.675 ] 00:16:27.675 }' 00:16:27.675 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:27.675 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:27.675 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:27.675 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:27.675 16:54:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:28.244 [2024-09-29 16:54:49.886559] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:28.244 [2024-09-29 16:54:49.886646] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:28.244 [2024-09-29 16:54:49.886761] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:28.503 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:28.503 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:28.503 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.762 "name": "raid_bdev1", 00:16:28.762 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:28.762 "strip_size_kb": 0, 00:16:28.762 "state": "online", 00:16:28.762 "raid_level": "raid1", 00:16:28.762 "superblock": true, 00:16:28.762 "num_base_bdevs": 2, 00:16:28.762 "num_base_bdevs_discovered": 2, 00:16:28.762 "num_base_bdevs_operational": 2, 00:16:28.762 "base_bdevs_list": [ 00:16:28.762 { 00:16:28.762 "name": "spare", 00:16:28.762 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:28.762 "is_configured": true, 00:16:28.762 "data_offset": 256, 00:16:28.762 "data_size": 7936 00:16:28.762 }, 00:16:28.762 { 00:16:28.762 "name": "BaseBdev2", 00:16:28.762 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:28.762 "is_configured": true, 00:16:28.762 "data_offset": 256, 00:16:28.762 "data_size": 7936 00:16:28.762 } 00:16:28.762 ] 00:16:28.762 }' 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:28.762 "name": "raid_bdev1", 00:16:28.762 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:28.762 "strip_size_kb": 0, 00:16:28.762 "state": "online", 00:16:28.762 "raid_level": "raid1", 00:16:28.762 "superblock": true, 00:16:28.762 "num_base_bdevs": 2, 00:16:28.762 "num_base_bdevs_discovered": 2, 00:16:28.762 "num_base_bdevs_operational": 2, 00:16:28.762 "base_bdevs_list": [ 00:16:28.762 { 00:16:28.762 "name": "spare", 00:16:28.762 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:28.762 "is_configured": true, 00:16:28.762 "data_offset": 256, 00:16:28.762 "data_size": 7936 00:16:28.762 }, 00:16:28.762 { 00:16:28.762 "name": "BaseBdev2", 00:16:28.762 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:28.762 "is_configured": true, 00:16:28.762 "data_offset": 256, 00:16:28.762 "data_size": 7936 00:16:28.762 } 00:16:28.762 ] 00:16:28.762 }' 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:28.762 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:29.020 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:29.021 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.021 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.021 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.021 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.021 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:29.021 "name": "raid_bdev1", 00:16:29.021 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:29.021 "strip_size_kb": 0, 00:16:29.021 "state": "online", 00:16:29.021 "raid_level": "raid1", 00:16:29.021 "superblock": true, 00:16:29.021 "num_base_bdevs": 2, 00:16:29.021 "num_base_bdevs_discovered": 2, 00:16:29.021 "num_base_bdevs_operational": 2, 00:16:29.021 "base_bdevs_list": [ 00:16:29.021 { 00:16:29.021 "name": "spare", 00:16:29.021 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:29.021 "is_configured": true, 00:16:29.021 "data_offset": 256, 00:16:29.021 "data_size": 7936 00:16:29.021 }, 00:16:29.021 { 00:16:29.021 "name": "BaseBdev2", 00:16:29.021 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:29.021 "is_configured": true, 00:16:29.021 "data_offset": 256, 00:16:29.021 "data_size": 7936 00:16:29.021 } 00:16:29.021 ] 00:16:29.021 }' 00:16:29.021 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:29.021 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.280 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:29.280 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.280 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.280 [2024-09-29 16:54:50.895984] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:29.280 [2024-09-29 16:54:50.896013] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:29.280 [2024-09-29 16:54:50.896084] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:29.280 [2024-09-29 16:54:50.896146] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:29.280 [2024-09-29 16:54:50.896160] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:29.280 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.280 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:29.281 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.542 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.542 [2024-09-29 16:54:50.959870] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:29.542 [2024-09-29 16:54:50.959923] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:29.542 [2024-09-29 16:54:50.959942] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:16:29.542 [2024-09-29 16:54:50.959952] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:29.542 [2024-09-29 16:54:50.961794] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:29.542 [2024-09-29 16:54:50.961830] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:29.542 [2024-09-29 16:54:50.961879] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:29.542 [2024-09-29 16:54:50.961934] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:29.542 [2024-09-29 16:54:50.962064] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:29.542 spare 00:16:29.542 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.542 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:29.542 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.542 16:54:50 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.542 [2024-09-29 16:54:51.061965] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:29.542 [2024-09-29 16:54:51.061996] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:29.542 [2024-09-29 16:54:51.062088] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:16:29.542 [2024-09-29 16:54:51.062158] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:29.542 [2024-09-29 16:54:51.062172] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:29.542 [2024-09-29 16:54:51.062249] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:29.542 "name": "raid_bdev1", 00:16:29.542 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:29.542 "strip_size_kb": 0, 00:16:29.542 "state": "online", 00:16:29.542 "raid_level": "raid1", 00:16:29.542 "superblock": true, 00:16:29.542 "num_base_bdevs": 2, 00:16:29.542 "num_base_bdevs_discovered": 2, 00:16:29.542 "num_base_bdevs_operational": 2, 00:16:29.542 "base_bdevs_list": [ 00:16:29.542 { 00:16:29.542 "name": "spare", 00:16:29.542 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:29.542 "is_configured": true, 00:16:29.542 "data_offset": 256, 00:16:29.542 "data_size": 7936 00:16:29.542 }, 00:16:29.542 { 00:16:29.542 "name": "BaseBdev2", 00:16:29.542 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:29.542 "is_configured": true, 00:16:29.542 "data_offset": 256, 00:16:29.542 "data_size": 7936 00:16:29.542 } 00:16:29.542 ] 00:16:29.542 }' 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:29.542 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:30.110 "name": "raid_bdev1", 00:16:30.110 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:30.110 "strip_size_kb": 0, 00:16:30.110 "state": "online", 00:16:30.110 "raid_level": "raid1", 00:16:30.110 "superblock": true, 00:16:30.110 "num_base_bdevs": 2, 00:16:30.110 "num_base_bdevs_discovered": 2, 00:16:30.110 "num_base_bdevs_operational": 2, 00:16:30.110 "base_bdevs_list": [ 00:16:30.110 { 00:16:30.110 "name": "spare", 00:16:30.110 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:30.110 "is_configured": true, 00:16:30.110 "data_offset": 256, 00:16:30.110 "data_size": 7936 00:16:30.110 }, 00:16:30.110 { 00:16:30.110 "name": "BaseBdev2", 00:16:30.110 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:30.110 "is_configured": true, 00:16:30.110 "data_offset": 256, 00:16:30.110 "data_size": 7936 00:16:30.110 } 00:16:30.110 ] 00:16:30.110 }' 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.110 [2024-09-29 16:54:51.654859] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.110 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:30.111 "name": "raid_bdev1", 00:16:30.111 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:30.111 "strip_size_kb": 0, 00:16:30.111 "state": "online", 00:16:30.111 "raid_level": "raid1", 00:16:30.111 "superblock": true, 00:16:30.111 "num_base_bdevs": 2, 00:16:30.111 "num_base_bdevs_discovered": 1, 00:16:30.111 "num_base_bdevs_operational": 1, 00:16:30.111 "base_bdevs_list": [ 00:16:30.111 { 00:16:30.111 "name": null, 00:16:30.111 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:30.111 "is_configured": false, 00:16:30.111 "data_offset": 0, 00:16:30.111 "data_size": 7936 00:16:30.111 }, 00:16:30.111 { 00:16:30.111 "name": "BaseBdev2", 00:16:30.111 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:30.111 "is_configured": true, 00:16:30.111 "data_offset": 256, 00:16:30.111 "data_size": 7936 00:16:30.111 } 00:16:30.111 ] 00:16:30.111 }' 00:16:30.111 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:30.111 16:54:51 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.679 16:54:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:30.679 16:54:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.679 16:54:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:30.679 [2024-09-29 16:54:52.074134] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:30.679 [2024-09-29 16:54:52.074260] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:30.679 [2024-09-29 16:54:52.074273] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:30.679 [2024-09-29 16:54:52.074311] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:30.679 [2024-09-29 16:54:52.077069] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:16:30.679 [2024-09-29 16:54:52.078901] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:30.679 16:54:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.679 16:54:52 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:31.617 "name": "raid_bdev1", 00:16:31.617 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:31.617 "strip_size_kb": 0, 00:16:31.617 "state": "online", 00:16:31.617 "raid_level": "raid1", 00:16:31.617 "superblock": true, 00:16:31.617 "num_base_bdevs": 2, 00:16:31.617 "num_base_bdevs_discovered": 2, 00:16:31.617 "num_base_bdevs_operational": 2, 00:16:31.617 "process": { 00:16:31.617 "type": "rebuild", 00:16:31.617 "target": "spare", 00:16:31.617 "progress": { 00:16:31.617 "blocks": 2560, 00:16:31.617 "percent": 32 00:16:31.617 } 00:16:31.617 }, 00:16:31.617 "base_bdevs_list": [ 00:16:31.617 { 00:16:31.617 "name": "spare", 00:16:31.617 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:31.617 "is_configured": true, 00:16:31.617 "data_offset": 256, 00:16:31.617 "data_size": 7936 00:16:31.617 }, 00:16:31.617 { 00:16:31.617 "name": "BaseBdev2", 00:16:31.617 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:31.617 "is_configured": true, 00:16:31.617 "data_offset": 256, 00:16:31.617 "data_size": 7936 00:16:31.617 } 00:16:31.617 ] 00:16:31.617 }' 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.617 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.617 [2024-09-29 16:54:53.221807] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:31.617 [2024-09-29 16:54:53.282893] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:31.617 [2024-09-29 16:54:53.282943] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:31.617 [2024-09-29 16:54:53.282959] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:31.617 [2024-09-29 16:54:53.282966] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.876 "name": "raid_bdev1", 00:16:31.876 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:31.876 "strip_size_kb": 0, 00:16:31.876 "state": "online", 00:16:31.876 "raid_level": "raid1", 00:16:31.876 "superblock": true, 00:16:31.876 "num_base_bdevs": 2, 00:16:31.876 "num_base_bdevs_discovered": 1, 00:16:31.876 "num_base_bdevs_operational": 1, 00:16:31.876 "base_bdevs_list": [ 00:16:31.876 { 00:16:31.876 "name": null, 00:16:31.876 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:31.876 "is_configured": false, 00:16:31.876 "data_offset": 0, 00:16:31.876 "data_size": 7936 00:16:31.876 }, 00:16:31.876 { 00:16:31.876 "name": "BaseBdev2", 00:16:31.876 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:31.876 "is_configured": true, 00:16:31.876 "data_offset": 256, 00:16:31.876 "data_size": 7936 00:16:31.876 } 00:16:31.876 ] 00:16:31.876 }' 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.876 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.135 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:32.135 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.135 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:32.135 [2024-09-29 16:54:53.745688] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:32.135 [2024-09-29 16:54:53.745751] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:32.135 [2024-09-29 16:54:53.745773] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:16:32.135 [2024-09-29 16:54:53.745782] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:32.135 [2024-09-29 16:54:53.745947] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:32.135 [2024-09-29 16:54:53.745959] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:32.135 [2024-09-29 16:54:53.746006] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:32.135 [2024-09-29 16:54:53.746016] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:32.135 [2024-09-29 16:54:53.746027] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:32.135 [2024-09-29 16:54:53.746047] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:32.135 [2024-09-29 16:54:53.748410] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:16:32.135 [2024-09-29 16:54:53.750285] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:32.135 spare 00:16:32.135 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.135 16:54:53 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.514 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:33.514 "name": "raid_bdev1", 00:16:33.514 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:33.514 "strip_size_kb": 0, 00:16:33.514 "state": "online", 00:16:33.514 "raid_level": "raid1", 00:16:33.514 "superblock": true, 00:16:33.514 "num_base_bdevs": 2, 00:16:33.514 "num_base_bdevs_discovered": 2, 00:16:33.514 "num_base_bdevs_operational": 2, 00:16:33.514 "process": { 00:16:33.514 "type": "rebuild", 00:16:33.514 "target": "spare", 00:16:33.514 "progress": { 00:16:33.514 "blocks": 2560, 00:16:33.514 "percent": 32 00:16:33.514 } 00:16:33.514 }, 00:16:33.514 "base_bdevs_list": [ 00:16:33.514 { 00:16:33.514 "name": "spare", 00:16:33.514 "uuid": "6a453a32-d67f-55e9-a2ac-690a47429d25", 00:16:33.514 "is_configured": true, 00:16:33.514 "data_offset": 256, 00:16:33.514 "data_size": 7936 00:16:33.514 }, 00:16:33.514 { 00:16:33.514 "name": "BaseBdev2", 00:16:33.514 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:33.514 "is_configured": true, 00:16:33.514 "data_offset": 256, 00:16:33.514 "data_size": 7936 00:16:33.514 } 00:16:33.514 ] 00:16:33.514 }' 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.515 [2024-09-29 16:54:54.889035] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:33.515 [2024-09-29 16:54:54.954249] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:33.515 [2024-09-29 16:54:54.954374] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:33.515 [2024-09-29 16:54:54.954389] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:33.515 [2024-09-29 16:54:54.954398] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.515 16:54:54 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.515 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:33.515 "name": "raid_bdev1", 00:16:33.515 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:33.515 "strip_size_kb": 0, 00:16:33.515 "state": "online", 00:16:33.515 "raid_level": "raid1", 00:16:33.515 "superblock": true, 00:16:33.515 "num_base_bdevs": 2, 00:16:33.515 "num_base_bdevs_discovered": 1, 00:16:33.515 "num_base_bdevs_operational": 1, 00:16:33.515 "base_bdevs_list": [ 00:16:33.515 { 00:16:33.515 "name": null, 00:16:33.515 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:33.515 "is_configured": false, 00:16:33.515 "data_offset": 0, 00:16:33.515 "data_size": 7936 00:16:33.515 }, 00:16:33.515 { 00:16:33.515 "name": "BaseBdev2", 00:16:33.515 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:33.515 "is_configured": true, 00:16:33.515 "data_offset": 256, 00:16:33.515 "data_size": 7936 00:16:33.515 } 00:16:33.515 ] 00:16:33.515 }' 00:16:33.515 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:33.515 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:33.774 "name": "raid_bdev1", 00:16:33.774 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:33.774 "strip_size_kb": 0, 00:16:33.774 "state": "online", 00:16:33.774 "raid_level": "raid1", 00:16:33.774 "superblock": true, 00:16:33.774 "num_base_bdevs": 2, 00:16:33.774 "num_base_bdevs_discovered": 1, 00:16:33.774 "num_base_bdevs_operational": 1, 00:16:33.774 "base_bdevs_list": [ 00:16:33.774 { 00:16:33.774 "name": null, 00:16:33.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:33.774 "is_configured": false, 00:16:33.774 "data_offset": 0, 00:16:33.774 "data_size": 7936 00:16:33.774 }, 00:16:33.774 { 00:16:33.774 "name": "BaseBdev2", 00:16:33.774 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:33.774 "is_configured": true, 00:16:33.774 "data_offset": 256, 00:16:33.774 "data_size": 7936 00:16:33.774 } 00:16:33.774 ] 00:16:33.774 }' 00:16:33.774 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:34.035 [2024-09-29 16:54:55.536567] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:34.035 [2024-09-29 16:54:55.536626] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.035 [2024-09-29 16:54:55.536646] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:16:34.035 [2024-09-29 16:54:55.536657] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.035 [2024-09-29 16:54:55.536821] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.035 [2024-09-29 16:54:55.536838] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:34.035 [2024-09-29 16:54:55.536881] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:34.035 [2024-09-29 16:54:55.536904] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:34.035 [2024-09-29 16:54:55.536910] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:34.035 [2024-09-29 16:54:55.536923] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:34.035 BaseBdev1 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.035 16:54:55 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:35.045 "name": "raid_bdev1", 00:16:35.045 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:35.045 "strip_size_kb": 0, 00:16:35.045 "state": "online", 00:16:35.045 "raid_level": "raid1", 00:16:35.045 "superblock": true, 00:16:35.045 "num_base_bdevs": 2, 00:16:35.045 "num_base_bdevs_discovered": 1, 00:16:35.045 "num_base_bdevs_operational": 1, 00:16:35.045 "base_bdevs_list": [ 00:16:35.045 { 00:16:35.045 "name": null, 00:16:35.045 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:35.045 "is_configured": false, 00:16:35.045 "data_offset": 0, 00:16:35.045 "data_size": 7936 00:16:35.045 }, 00:16:35.045 { 00:16:35.045 "name": "BaseBdev2", 00:16:35.045 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:35.045 "is_configured": true, 00:16:35.045 "data_offset": 256, 00:16:35.045 "data_size": 7936 00:16:35.045 } 00:16:35.045 ] 00:16:35.045 }' 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:35.045 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.304 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:35.304 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:35.304 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:35.305 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:35.305 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:35.305 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:35.305 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.305 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.305 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:35.565 16:54:56 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:35.565 "name": "raid_bdev1", 00:16:35.565 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:35.565 "strip_size_kb": 0, 00:16:35.565 "state": "online", 00:16:35.565 "raid_level": "raid1", 00:16:35.565 "superblock": true, 00:16:35.565 "num_base_bdevs": 2, 00:16:35.565 "num_base_bdevs_discovered": 1, 00:16:35.565 "num_base_bdevs_operational": 1, 00:16:35.565 "base_bdevs_list": [ 00:16:35.565 { 00:16:35.565 "name": null, 00:16:35.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:35.565 "is_configured": false, 00:16:35.565 "data_offset": 0, 00:16:35.565 "data_size": 7936 00:16:35.565 }, 00:16:35.565 { 00:16:35.565 "name": "BaseBdev2", 00:16:35.565 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:35.565 "is_configured": true, 00:16:35.565 "data_offset": 256, 00:16:35.565 "data_size": 7936 00:16:35.565 } 00:16:35.565 ] 00:16:35.565 }' 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:35.565 [2024-09-29 16:54:57.101865] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:35.565 [2024-09-29 16:54:57.101998] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:35.565 [2024-09-29 16:54:57.102010] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:35.565 request: 00:16:35.565 { 00:16:35.565 "base_bdev": "BaseBdev1", 00:16:35.565 "raid_bdev": "raid_bdev1", 00:16:35.565 "method": "bdev_raid_add_base_bdev", 00:16:35.565 "req_id": 1 00:16:35.565 } 00:16:35.565 Got JSON-RPC error response 00:16:35.565 response: 00:16:35.565 { 00:16:35.565 "code": -22, 00:16:35.565 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:35.565 } 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:35.565 16:54:57 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:36.505 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:36.505 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:36.505 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:36.505 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:36.505 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:36.505 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:36.505 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:36.506 "name": "raid_bdev1", 00:16:36.506 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:36.506 "strip_size_kb": 0, 00:16:36.506 "state": "online", 00:16:36.506 "raid_level": "raid1", 00:16:36.506 "superblock": true, 00:16:36.506 "num_base_bdevs": 2, 00:16:36.506 "num_base_bdevs_discovered": 1, 00:16:36.506 "num_base_bdevs_operational": 1, 00:16:36.506 "base_bdevs_list": [ 00:16:36.506 { 00:16:36.506 "name": null, 00:16:36.506 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:36.506 "is_configured": false, 00:16:36.506 "data_offset": 0, 00:16:36.506 "data_size": 7936 00:16:36.506 }, 00:16:36.506 { 00:16:36.506 "name": "BaseBdev2", 00:16:36.506 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:36.506 "is_configured": true, 00:16:36.506 "data_offset": 256, 00:16:36.506 "data_size": 7936 00:16:36.506 } 00:16:36.506 ] 00:16:36.506 }' 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:36.506 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:37.075 "name": "raid_bdev1", 00:16:37.075 "uuid": "21b8f84e-ca24-4a7f-a2df-2b9b4d21baa6", 00:16:37.075 "strip_size_kb": 0, 00:16:37.075 "state": "online", 00:16:37.075 "raid_level": "raid1", 00:16:37.075 "superblock": true, 00:16:37.075 "num_base_bdevs": 2, 00:16:37.075 "num_base_bdevs_discovered": 1, 00:16:37.075 "num_base_bdevs_operational": 1, 00:16:37.075 "base_bdevs_list": [ 00:16:37.075 { 00:16:37.075 "name": null, 00:16:37.075 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:37.075 "is_configured": false, 00:16:37.075 "data_offset": 0, 00:16:37.075 "data_size": 7936 00:16:37.075 }, 00:16:37.075 { 00:16:37.075 "name": "BaseBdev2", 00:16:37.075 "uuid": "cc83bb5f-ea9d-5864-beb6-6671a6fddd79", 00:16:37.075 "is_configured": true, 00:16:37.075 "data_offset": 256, 00:16:37.075 "data_size": 7936 00:16:37.075 } 00:16:37.075 ] 00:16:37.075 }' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 98961 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 98961 ']' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 98961 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 98961 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:37.075 killing process with pid 98961 00:16:37.075 Received shutdown signal, test time was about 60.000000 seconds 00:16:37.075 00:16:37.075 Latency(us) 00:16:37.075 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:37.075 =================================================================================================================== 00:16:37.075 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 98961' 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 98961 00:16:37.075 [2024-09-29 16:54:58.743535] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:37.075 [2024-09-29 16:54:58.743639] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:37.075 [2024-09-29 16:54:58.743682] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:37.075 [2024-09-29 16:54:58.743692] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:37.075 16:54:58 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 98961 00:16:37.336 [2024-09-29 16:54:58.777544] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:37.336 16:54:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:16:37.336 00:16:37.336 real 0m16.026s 00:16:37.336 user 0m21.357s 00:16:37.336 sys 0m1.646s 00:16:37.336 16:54:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:37.336 ************************************ 00:16:37.336 END TEST raid_rebuild_test_sb_md_interleaved 00:16:37.336 ************************************ 00:16:37.336 16:54:59 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:37.595 16:54:59 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:16:37.595 16:54:59 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:16:37.595 16:54:59 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 98961 ']' 00:16:37.595 16:54:59 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 98961 00:16:37.595 16:54:59 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:16:37.595 ************************************ 00:16:37.595 END TEST bdev_raid 00:16:37.595 ************************************ 00:16:37.595 00:16:37.595 real 9m53.711s 00:16:37.595 user 14m5.555s 00:16:37.595 sys 1m47.244s 00:16:37.595 16:54:59 bdev_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:37.595 16:54:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:37.595 16:54:59 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:37.595 16:54:59 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:16:37.595 16:54:59 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:37.595 16:54:59 -- common/autotest_common.sh@10 -- # set +x 00:16:37.595 ************************************ 00:16:37.595 START TEST spdkcli_raid 00:16:37.595 ************************************ 00:16:37.595 16:54:59 spdkcli_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:37.854 * Looking for test storage... 00:16:37.854 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:37.854 16:54:59 spdkcli_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:16:37.854 16:54:59 spdkcli_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:16:37.854 16:54:59 spdkcli_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:16:37.854 16:54:59 spdkcli_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:16:37.854 16:54:59 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:16:37.855 16:54:59 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:37.855 16:54:59 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:16:37.855 16:54:59 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:16:37.855 16:54:59 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:37.855 16:54:59 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:37.855 16:54:59 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:16:37.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:37.855 --rc genhtml_branch_coverage=1 00:16:37.855 --rc genhtml_function_coverage=1 00:16:37.855 --rc genhtml_legend=1 00:16:37.855 --rc geninfo_all_blocks=1 00:16:37.855 --rc geninfo_unexecuted_blocks=1 00:16:37.855 00:16:37.855 ' 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:16:37.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:37.855 --rc genhtml_branch_coverage=1 00:16:37.855 --rc genhtml_function_coverage=1 00:16:37.855 --rc genhtml_legend=1 00:16:37.855 --rc geninfo_all_blocks=1 00:16:37.855 --rc geninfo_unexecuted_blocks=1 00:16:37.855 00:16:37.855 ' 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:16:37.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:37.855 --rc genhtml_branch_coverage=1 00:16:37.855 --rc genhtml_function_coverage=1 00:16:37.855 --rc genhtml_legend=1 00:16:37.855 --rc geninfo_all_blocks=1 00:16:37.855 --rc geninfo_unexecuted_blocks=1 00:16:37.855 00:16:37.855 ' 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:16:37.855 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:37.855 --rc genhtml_branch_coverage=1 00:16:37.855 --rc genhtml_function_coverage=1 00:16:37.855 --rc genhtml_legend=1 00:16:37.855 --rc geninfo_all_blocks=1 00:16:37.855 --rc geninfo_unexecuted_blocks=1 00:16:37.855 00:16:37.855 ' 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:16:37.855 16:54:59 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=99631 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:16:37.855 16:54:59 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 99631 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@831 -- # '[' -z 99631 ']' 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:37.855 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:37.855 16:54:59 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:38.114 [2024-09-29 16:54:59.553341] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:38.114 [2024-09-29 16:54:59.553493] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99631 ] 00:16:38.114 [2024-09-29 16:54:59.701396] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:16:38.114 [2024-09-29 16:54:59.749667] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:38.114 [2024-09-29 16:54:59.749790] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:16:38.684 16:55:00 spdkcli_raid -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:38.684 16:55:00 spdkcli_raid -- common/autotest_common.sh@864 -- # return 0 00:16:38.684 16:55:00 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:16:38.684 16:55:00 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:38.684 16:55:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:38.943 16:55:00 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:16:38.943 16:55:00 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:38.943 16:55:00 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:38.943 16:55:00 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:16:38.943 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:16:38.943 ' 00:16:40.324 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:16:40.324 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:16:40.324 16:55:01 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:16:40.324 16:55:01 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:40.324 16:55:01 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:40.584 16:55:02 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:16:40.584 16:55:02 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:40.584 16:55:02 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:40.584 16:55:02 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:16:40.584 ' 00:16:41.524 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:16:41.784 16:55:03 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:16:41.784 16:55:03 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:41.784 16:55:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:41.784 16:55:03 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:16:41.784 16:55:03 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:41.784 16:55:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:41.784 16:55:03 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:16:41.784 16:55:03 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:16:42.354 16:55:03 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:16:42.354 16:55:03 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:16:42.354 16:55:03 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:16:42.354 16:55:03 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:42.354 16:55:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:42.354 16:55:03 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:16:42.354 16:55:03 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:42.354 16:55:03 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:42.354 16:55:03 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:16:42.354 ' 00:16:43.293 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:16:43.293 16:55:04 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:16:43.293 16:55:04 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:43.293 16:55:04 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:43.553 16:55:04 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:16:43.553 16:55:04 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:43.553 16:55:04 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:43.553 16:55:04 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:16:43.553 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:16:43.553 ' 00:16:44.935 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:16:44.935 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:16:44.935 16:55:06 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:44.935 16:55:06 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 99631 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 99631 ']' 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 99631 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@955 -- # uname 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 99631 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@968 -- # echo 'killing process with pid 99631' 00:16:44.935 killing process with pid 99631 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@969 -- # kill 99631 00:16:44.935 16:55:06 spdkcli_raid -- common/autotest_common.sh@974 -- # wait 99631 00:16:45.505 16:55:06 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:16:45.505 16:55:06 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 99631 ']' 00:16:45.505 16:55:06 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 99631 00:16:45.505 16:55:06 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 99631 ']' 00:16:45.505 16:55:06 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 99631 00:16:45.505 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (99631) - No such process 00:16:45.505 16:55:06 spdkcli_raid -- common/autotest_common.sh@977 -- # echo 'Process with pid 99631 is not found' 00:16:45.505 Process with pid 99631 is not found 00:16:45.505 16:55:06 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:16:45.505 16:55:06 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:16:45.505 16:55:06 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:16:45.505 16:55:06 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:16:45.505 00:16:45.505 real 0m7.760s 00:16:45.505 user 0m16.300s 00:16:45.505 sys 0m1.124s 00:16:45.505 16:55:06 spdkcli_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:45.505 16:55:06 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:16:45.505 ************************************ 00:16:45.505 END TEST spdkcli_raid 00:16:45.505 ************************************ 00:16:45.505 16:55:06 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:45.505 16:55:06 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:45.505 16:55:06 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:45.505 16:55:06 -- common/autotest_common.sh@10 -- # set +x 00:16:45.505 ************************************ 00:16:45.505 START TEST blockdev_raid5f 00:16:45.505 ************************************ 00:16:45.505 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:16:45.505 * Looking for test storage... 00:16:45.505 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:16:45.505 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:16:45.505 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lcov --version 00:16:45.505 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:45.766 16:55:07 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:16:45.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:45.766 --rc genhtml_branch_coverage=1 00:16:45.766 --rc genhtml_function_coverage=1 00:16:45.766 --rc genhtml_legend=1 00:16:45.766 --rc geninfo_all_blocks=1 00:16:45.766 --rc geninfo_unexecuted_blocks=1 00:16:45.766 00:16:45.766 ' 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:16:45.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:45.766 --rc genhtml_branch_coverage=1 00:16:45.766 --rc genhtml_function_coverage=1 00:16:45.766 --rc genhtml_legend=1 00:16:45.766 --rc geninfo_all_blocks=1 00:16:45.766 --rc geninfo_unexecuted_blocks=1 00:16:45.766 00:16:45.766 ' 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:16:45.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:45.766 --rc genhtml_branch_coverage=1 00:16:45.766 --rc genhtml_function_coverage=1 00:16:45.766 --rc genhtml_legend=1 00:16:45.766 --rc geninfo_all_blocks=1 00:16:45.766 --rc geninfo_unexecuted_blocks=1 00:16:45.766 00:16:45.766 ' 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:16:45.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:45.766 --rc genhtml_branch_coverage=1 00:16:45.766 --rc genhtml_function_coverage=1 00:16:45.766 --rc genhtml_legend=1 00:16:45.766 --rc geninfo_all_blocks=1 00:16:45.766 --rc geninfo_unexecuted_blocks=1 00:16:45.766 00:16:45.766 ' 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@669 -- # QOS_DEV_1=Malloc_0 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@670 -- # QOS_DEV_2=Null_1 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@671 -- # QOS_RUN_TIME=5 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@673 -- # uname -s 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@673 -- # '[' Linux = Linux ']' 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@675 -- # PRE_RESERVED_MEM=0 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@681 -- # test_type=raid5f 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@682 -- # crypto_device= 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@683 -- # dek= 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@684 -- # env_ctx= 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@685 -- # wait_for_rpc= 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@686 -- # '[' -n '' ']' 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == bdev ]] 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == crypto_* ]] 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@692 -- # start_spdk_tgt 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=99886 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:16:45.766 16:55:07 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 99886 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@831 -- # '[' -z 99886 ']' 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:45.766 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:45.766 16:55:07 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:45.766 [2024-09-29 16:55:07.354091] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:45.766 [2024-09-29 16:55:07.354310] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99886 ] 00:16:46.026 [2024-09-29 16:55:07.495064] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:46.026 [2024-09-29 16:55:07.541468] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@864 -- # return 0 00:16:46.597 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@693 -- # case "$test_type" in 00:16:46.597 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@725 -- # setup_raid5f_conf 00:16:46.597 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:46.597 Malloc0 00:16:46.597 Malloc1 00:16:46.597 Malloc2 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.597 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@736 -- # rpc_cmd bdev_wait_for_examine 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.597 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@739 -- # cat 00:16:46.597 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n accel 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.597 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n bdev 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.597 16:55:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n iobuf 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@747 -- # mapfile -t bdevs 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@747 -- # rpc_cmd bdev_get_bdevs 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@747 -- # jq -r '.[] | select(.claimed == false)' 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@748 -- # mapfile -t bdevs_name 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@748 -- # jq -r .name 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@748 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "a1063c81-aaba-47ba-a536-c61e3a6a8ab1"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "a1063c81-aaba-47ba-a536-c61e3a6a8ab1",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "a1063c81-aaba-47ba-a536-c61e3a6a8ab1",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "e11f2eb0-c9e6-4b97-a071-a8f18c7eceec",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "eaa43389-3a46-4660-9e1a-d0271ade46b2",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "f1f17a1b-dab5-4060-b5e5-cdd04fe5c274",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@749 -- # bdev_list=("${bdevs_name[@]}") 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@751 -- # hello_world_bdev=raid5f 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@752 -- # trap - SIGINT SIGTERM EXIT 00:16:46.858 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@753 -- # killprocess 99886 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@950 -- # '[' -z 99886 ']' 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@954 -- # kill -0 99886 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@955 -- # uname 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 99886 00:16:46.858 killing process with pid 99886 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@968 -- # echo 'killing process with pid 99886' 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@969 -- # kill 99886 00:16:46.858 16:55:08 blockdev_raid5f -- common/autotest_common.sh@974 -- # wait 99886 00:16:47.428 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@757 -- # trap cleanup SIGINT SIGTERM EXIT 00:16:47.428 16:55:08 blockdev_raid5f -- bdev/blockdev.sh@759 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:47.428 16:55:08 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:16:47.428 16:55:08 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:47.428 16:55:08 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:47.428 ************************************ 00:16:47.428 START TEST bdev_hello_world 00:16:47.428 ************************************ 00:16:47.428 16:55:08 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:16:47.428 [2024-09-29 16:55:08.979626] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:47.428 [2024-09-29 16:55:08.979753] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99926 ] 00:16:47.688 [2024-09-29 16:55:09.124460] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:47.688 [2024-09-29 16:55:09.171817] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:47.948 [2024-09-29 16:55:09.368301] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:16:47.948 [2024-09-29 16:55:09.368362] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:16:47.948 [2024-09-29 16:55:09.368384] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:16:47.948 [2024-09-29 16:55:09.368670] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:16:47.948 [2024-09-29 16:55:09.368823] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:16:47.948 [2024-09-29 16:55:09.368842] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:16:47.948 [2024-09-29 16:55:09.368905] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:16:47.948 00:16:47.948 [2024-09-29 16:55:09.368921] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:16:48.209 00:16:48.209 real 0m0.727s 00:16:48.209 user 0m0.392s 00:16:48.209 sys 0m0.217s 00:16:48.209 16:55:09 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:48.209 ************************************ 00:16:48.209 END TEST bdev_hello_world 00:16:48.209 ************************************ 00:16:48.209 16:55:09 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:16:48.209 16:55:09 blockdev_raid5f -- bdev/blockdev.sh@760 -- # run_test bdev_bounds bdev_bounds '' 00:16:48.209 16:55:09 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:48.209 16:55:09 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:48.209 16:55:09 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:48.209 ************************************ 00:16:48.209 START TEST bdev_bounds 00:16:48.209 ************************************ 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1125 -- # bdev_bounds '' 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=99957 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:16:48.209 Process bdevio pid: 99957 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 99957' 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 99957 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@831 -- # '[' -z 99957 ']' 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:48.209 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:48.209 16:55:09 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:48.209 [2024-09-29 16:55:09.783884] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:48.209 [2024-09-29 16:55:09.784078] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99957 ] 00:16:48.469 [2024-09-29 16:55:09.930671] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:16:48.469 [2024-09-29 16:55:09.976950] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:16:48.469 [2024-09-29 16:55:09.977052] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:48.469 [2024-09-29 16:55:09.977160] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:16:49.039 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:49.039 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # return 0 00:16:49.039 16:55:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:16:49.039 I/O targets: 00:16:49.039 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:16:49.039 00:16:49.039 00:16:49.039 CUnit - A unit testing framework for C - Version 2.1-3 00:16:49.039 http://cunit.sourceforge.net/ 00:16:49.039 00:16:49.039 00:16:49.039 Suite: bdevio tests on: raid5f 00:16:49.039 Test: blockdev write read block ...passed 00:16:49.039 Test: blockdev write zeroes read block ...passed 00:16:49.299 Test: blockdev write zeroes read no split ...passed 00:16:49.299 Test: blockdev write zeroes read split ...passed 00:16:49.299 Test: blockdev write zeroes read split partial ...passed 00:16:49.299 Test: blockdev reset ...passed 00:16:49.299 Test: blockdev write read 8 blocks ...passed 00:16:49.299 Test: blockdev write read size > 128k ...passed 00:16:49.299 Test: blockdev write read invalid size ...passed 00:16:49.299 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:16:49.299 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:16:49.299 Test: blockdev write read max offset ...passed 00:16:49.299 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:16:49.299 Test: blockdev writev readv 8 blocks ...passed 00:16:49.299 Test: blockdev writev readv 30 x 1block ...passed 00:16:49.299 Test: blockdev writev readv block ...passed 00:16:49.299 Test: blockdev writev readv size > 128k ...passed 00:16:49.299 Test: blockdev writev readv size > 128k in two iovs ...passed 00:16:49.299 Test: blockdev comparev and writev ...passed 00:16:49.299 Test: blockdev nvme passthru rw ...passed 00:16:49.299 Test: blockdev nvme passthru vendor specific ...passed 00:16:49.299 Test: blockdev nvme admin passthru ...passed 00:16:49.299 Test: blockdev copy ...passed 00:16:49.299 00:16:49.299 Run Summary: Type Total Ran Passed Failed Inactive 00:16:49.299 suites 1 1 n/a 0 0 00:16:49.299 tests 23 23 23 0 0 00:16:49.299 asserts 130 130 130 0 n/a 00:16:49.299 00:16:49.299 Elapsed time = 0.321 seconds 00:16:49.299 0 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 99957 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@950 -- # '[' -z 99957 ']' 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # kill -0 99957 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # uname 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 99957 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@968 -- # echo 'killing process with pid 99957' 00:16:49.299 killing process with pid 99957 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@969 -- # kill 99957 00:16:49.299 16:55:10 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@974 -- # wait 99957 00:16:49.560 16:55:11 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:16:49.560 00:16:49.560 real 0m1.496s 00:16:49.560 user 0m3.583s 00:16:49.560 sys 0m0.355s 00:16:49.560 16:55:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:49.560 16:55:11 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:16:49.560 ************************************ 00:16:49.560 END TEST bdev_bounds 00:16:49.560 ************************************ 00:16:49.821 16:55:11 blockdev_raid5f -- bdev/blockdev.sh@761 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:49.821 16:55:11 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:16:49.821 16:55:11 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:49.821 16:55:11 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:49.821 ************************************ 00:16:49.821 START TEST bdev_nbd 00:16:49.821 ************************************ 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1125 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:16:49.821 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=100001 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 100001 /var/tmp/spdk-nbd.sock 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@831 -- # '[' -z 100001 ']' 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:49.821 16:55:11 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:49.821 [2024-09-29 16:55:11.383998] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:16:49.821 [2024-09-29 16:55:11.384251] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:50.082 [2024-09-29 16:55:11.533069] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:50.082 [2024-09-29 16:55:11.578715] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:50.652 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # return 0 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:50.653 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:50.919 1+0 records in 00:16:50.919 1+0 records out 00:16:50.919 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000540838 s, 7.6 MB/s 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:16:50.919 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:16:51.179 { 00:16:51.179 "nbd_device": "/dev/nbd0", 00:16:51.179 "bdev_name": "raid5f" 00:16:51.179 } 00:16:51.179 ]' 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:16:51.179 { 00:16:51.179 "nbd_device": "/dev/nbd0", 00:16:51.179 "bdev_name": "raid5f" 00:16:51.179 } 00:16:51.179 ]' 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:51.179 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:51.439 16:55:12 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:51.699 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:16:51.959 /dev/nbd0 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:51.959 1+0 records in 00:16:51.959 1+0 records out 00:16:51.959 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000450619 s, 9.1 MB/s 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:51.959 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:16:52.218 { 00:16:52.218 "nbd_device": "/dev/nbd0", 00:16:52.218 "bdev_name": "raid5f" 00:16:52.218 } 00:16:52.218 ]' 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:16:52.218 { 00:16:52.218 "nbd_device": "/dev/nbd0", 00:16:52.218 "bdev_name": "raid5f" 00:16:52.218 } 00:16:52.218 ]' 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:16:52.218 256+0 records in 00:16:52.218 256+0 records out 00:16:52.218 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0121313 s, 86.4 MB/s 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:16:52.218 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:16:52.218 256+0 records in 00:16:52.218 256+0 records out 00:16:52.219 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0288162 s, 36.4 MB/s 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:52.219 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:52.477 16:55:13 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:16:52.736 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:16:52.996 malloc_lvol_verify 00:16:52.996 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:16:52.996 b27a04d8-44e2-4337-84aa-71a7e2712286 00:16:52.996 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:16:53.255 7bbdfed7-029d-4cba-9e05-d0f529e7e571 00:16:53.255 16:55:14 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:16:53.513 /dev/nbd0 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:16:53.513 mke2fs 1.47.0 (5-Feb-2023) 00:16:53.513 Discarding device blocks: 0/4096 done 00:16:53.513 Creating filesystem with 4096 1k blocks and 1024 inodes 00:16:53.513 00:16:53.513 Allocating group tables: 0/1 done 00:16:53.513 Writing inode tables: 0/1 done 00:16:53.513 Creating journal (1024 blocks): done 00:16:53.513 Writing superblocks and filesystem accounting information: 0/1 done 00:16:53.513 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:53.513 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 100001 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@950 -- # '[' -z 100001 ']' 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # kill -0 100001 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # uname 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 100001 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:53.773 killing process with pid 100001 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@968 -- # echo 'killing process with pid 100001' 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@969 -- # kill 100001 00:16:53.773 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@974 -- # wait 100001 00:16:54.033 16:55:15 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:16:54.033 00:16:54.033 real 0m4.303s 00:16:54.033 user 0m6.258s 00:16:54.033 sys 0m1.224s 00:16:54.033 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:54.033 16:55:15 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:16:54.033 ************************************ 00:16:54.033 END TEST bdev_nbd 00:16:54.033 ************************************ 00:16:54.033 16:55:15 blockdev_raid5f -- bdev/blockdev.sh@762 -- # [[ y == y ]] 00:16:54.033 16:55:15 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = nvme ']' 00:16:54.033 16:55:15 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = gpt ']' 00:16:54.033 16:55:15 blockdev_raid5f -- bdev/blockdev.sh@767 -- # run_test bdev_fio fio_test_suite '' 00:16:54.033 16:55:15 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:54.033 16:55:15 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:54.033 16:55:15 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:16:54.033 ************************************ 00:16:54.033 START TEST bdev_fio 00:16:54.033 ************************************ 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1125 -- # fio_test_suite '' 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:16:54.033 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:16:54.033 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:16:54.034 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:16:54.034 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1101 -- # '[' 11 -le 1 ']' 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:16:54.294 ************************************ 00:16:54.294 START TEST bdev_fio_rw_verify 00:16:54.294 ************************************ 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1125 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib=/usr/lib64/libasan.so.8 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # break 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:16:54.294 16:55:15 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:16:54.553 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:16:54.553 fio-3.35 00:16:54.553 Starting 1 thread 00:17:06.770 00:17:06.770 job_raid5f: (groupid=0, jobs=1): err= 0: pid=100196: Sun Sep 29 16:55:26 2024 00:17:06.770 read: IOPS=12.2k, BW=47.8MiB/s (50.1MB/s)(478MiB/10001msec) 00:17:06.770 slat (usec): min=16, max=174, avg=19.21, stdev= 1.99 00:17:06.770 clat (usec): min=11, max=492, avg=130.79, stdev=45.69 00:17:06.770 lat (usec): min=31, max=512, avg=150.00, stdev=45.97 00:17:06.770 clat percentiles (usec): 00:17:06.770 | 50.000th=[ 135], 99.000th=[ 217], 99.900th=[ 235], 99.990th=[ 281], 00:17:06.770 | 99.999th=[ 474] 00:17:06.770 write: IOPS=12.8k, BW=50.1MiB/s (52.5MB/s)(494MiB/9874msec); 0 zone resets 00:17:06.770 slat (usec): min=8, max=324, avg=16.92, stdev= 4.62 00:17:06.770 clat (usec): min=60, max=1735, avg=300.81, stdev=53.01 00:17:06.770 lat (usec): min=76, max=1969, avg=317.73, stdev=54.93 00:17:06.770 clat percentiles (usec): 00:17:06.770 | 50.000th=[ 306], 99.000th=[ 383], 99.900th=[ 930], 99.990th=[ 1565], 00:17:06.770 | 99.999th=[ 1713] 00:17:06.770 bw ( KiB/s): min=47920, max=53608, per=98.93%, avg=50716.42, stdev=1437.27, samples=19 00:17:06.770 iops : min=11980, max=13402, avg=12679.11, stdev=359.32, samples=19 00:17:06.770 lat (usec) : 20=0.01%, 50=0.01%, 100=16.20%, 250=39.34%, 500=44.29% 00:17:06.770 lat (usec) : 750=0.09%, 1000=0.04% 00:17:06.770 lat (msec) : 2=0.04% 00:17:06.770 cpu : usr=98.70%, sys=0.50%, ctx=25, majf=0, minf=13110 00:17:06.770 IO depths : 1=7.6%, 2=19.8%, 4=55.2%, 8=17.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:06.770 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:06.770 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:06.770 issued rwts: total=122379,126551,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:06.770 latency : target=0, window=0, percentile=100.00%, depth=8 00:17:06.770 00:17:06.770 Run status group 0 (all jobs): 00:17:06.770 READ: bw=47.8MiB/s (50.1MB/s), 47.8MiB/s-47.8MiB/s (50.1MB/s-50.1MB/s), io=478MiB (501MB), run=10001-10001msec 00:17:06.770 WRITE: bw=50.1MiB/s (52.5MB/s), 50.1MiB/s-50.1MiB/s (52.5MB/s-52.5MB/s), io=494MiB (518MB), run=9874-9874msec 00:17:06.770 ----------------------------------------------------- 00:17:06.770 Suppressions used: 00:17:06.770 count bytes template 00:17:06.770 1 7 /usr/src/fio/parse.c 00:17:06.770 383 36768 /usr/src/fio/iolog.c 00:17:06.770 1 8 libtcmalloc_minimal.so 00:17:06.770 1 904 libcrypto.so 00:17:06.770 ----------------------------------------------------- 00:17:06.770 00:17:06.770 00:17:06.770 real 0m11.201s 00:17:06.770 user 0m11.523s 00:17:06.770 sys 0m0.677s 00:17:06.770 16:55:27 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:06.770 16:55:27 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:17:06.770 ************************************ 00:17:06.770 END TEST bdev_fio_rw_verify 00:17:06.770 ************************************ 00:17:06.770 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:17:06.770 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:06.770 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "a1063c81-aaba-47ba-a536-c61e3a6a8ab1"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "a1063c81-aaba-47ba-a536-c61e3a6a8ab1",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "a1063c81-aaba-47ba-a536-c61e3a6a8ab1",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "e11f2eb0-c9e6-4b97-a071-a8f18c7eceec",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "eaa43389-3a46-4660-9e1a-d0271ade46b2",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "f1f17a1b-dab5-4060-b5e5-cdd04fe5c274",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:06.771 /home/vagrant/spdk_repo/spdk 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:17:06.771 00:17:06.771 real 0m11.485s 00:17:06.771 user 0m11.646s 00:17:06.771 sys 0m0.814s 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:06.771 16:55:27 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:06.771 ************************************ 00:17:06.771 END TEST bdev_fio 00:17:06.771 ************************************ 00:17:06.771 16:55:27 blockdev_raid5f -- bdev/blockdev.sh@774 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:06.771 16:55:27 blockdev_raid5f -- bdev/blockdev.sh@776 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:06.771 16:55:27 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:17:06.771 16:55:27 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:06.771 16:55:27 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:06.771 ************************************ 00:17:06.771 START TEST bdev_verify 00:17:06.771 ************************************ 00:17:06.771 16:55:27 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:06.771 [2024-09-29 16:55:27.272339] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:17:06.771 [2024-09-29 16:55:27.272442] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100353 ] 00:17:06.771 [2024-09-29 16:55:27.397565] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:06.771 [2024-09-29 16:55:27.442292] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:06.771 [2024-09-29 16:55:27.442378] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:06.771 Running I/O for 5 seconds... 00:17:11.268 11070.00 IOPS, 43.24 MiB/s 11181.00 IOPS, 43.68 MiB/s 11183.00 IOPS, 43.68 MiB/s 11186.00 IOPS, 43.70 MiB/s 11210.20 IOPS, 43.79 MiB/s 00:17:11.268 Latency(us) 00:17:11.268 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:11.268 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:17:11.268 Verification LBA range: start 0x0 length 0x2000 00:17:11.268 raid5f : 5.02 4477.99 17.49 0.00 0.00 42782.70 133.25 31136.75 00:17:11.268 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:17:11.268 Verification LBA range: start 0x2000 length 0x2000 00:17:11.268 raid5f : 5.01 6721.04 26.25 0.00 0.00 28577.49 320.17 21292.05 00:17:11.268 =================================================================================================================== 00:17:11.268 Total : 11199.03 43.75 0.00 0.00 34264.73 133.25 31136.75 00:17:11.268 00:17:11.268 real 0m5.714s 00:17:11.268 user 0m10.664s 00:17:11.268 sys 0m0.228s 00:17:11.268 16:55:32 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:11.268 16:55:32 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:17:11.268 ************************************ 00:17:11.268 END TEST bdev_verify 00:17:11.268 ************************************ 00:17:11.528 16:55:32 blockdev_raid5f -- bdev/blockdev.sh@777 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:11.528 16:55:32 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:17:11.528 16:55:32 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:11.528 16:55:32 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:11.528 ************************************ 00:17:11.528 START TEST bdev_verify_big_io 00:17:11.528 ************************************ 00:17:11.528 16:55:32 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:11.528 [2024-09-29 16:55:33.068333] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:17:11.528 [2024-09-29 16:55:33.068481] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100430 ] 00:17:11.789 [2024-09-29 16:55:33.215459] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:11.789 [2024-09-29 16:55:33.271036] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:11.789 [2024-09-29 16:55:33.271134] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:12.049 Running I/O for 5 seconds... 00:17:17.193 633.00 IOPS, 39.56 MiB/s 761.00 IOPS, 47.56 MiB/s 803.67 IOPS, 50.23 MiB/s 824.50 IOPS, 51.53 MiB/s 812.40 IOPS, 50.77 MiB/s 00:17:17.193 Latency(us) 00:17:17.193 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:17.193 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:17:17.193 Verification LBA range: start 0x0 length 0x200 00:17:17.193 raid5f : 5.28 360.96 22.56 0.00 0.00 8754796.23 201.22 380967.35 00:17:17.193 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:17:17.193 Verification LBA range: start 0x200 length 0x200 00:17:17.193 raid5f : 5.19 464.98 29.06 0.00 0.00 6876724.56 182.44 296714.96 00:17:17.193 =================================================================================================================== 00:17:17.193 Total : 825.94 51.62 0.00 0.00 7705285.59 182.44 380967.35 00:17:17.454 00:17:17.454 real 0m6.018s 00:17:17.454 user 0m11.194s 00:17:17.454 sys 0m0.260s 00:17:17.454 16:55:38 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:17.454 16:55:38 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:17:17.454 ************************************ 00:17:17.454 END TEST bdev_verify_big_io 00:17:17.454 ************************************ 00:17:17.454 16:55:39 blockdev_raid5f -- bdev/blockdev.sh@778 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:17.454 16:55:39 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:17:17.454 16:55:39 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:17.454 16:55:39 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:17.454 ************************************ 00:17:17.454 START TEST bdev_write_zeroes 00:17:17.454 ************************************ 00:17:17.454 16:55:39 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:17.713 [2024-09-29 16:55:39.167185] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:17:17.713 [2024-09-29 16:55:39.167321] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100512 ] 00:17:17.713 [2024-09-29 16:55:39.313487] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:17.713 [2024-09-29 16:55:39.370421] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:17.973 Running I/O for 1 seconds... 00:17:18.912 30255.00 IOPS, 118.18 MiB/s 00:17:18.912 Latency(us) 00:17:18.912 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:18.912 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:17:18.912 raid5f : 1.01 30222.67 118.06 0.00 0.00 4220.39 1359.37 5752.29 00:17:18.912 =================================================================================================================== 00:17:18.912 Total : 30222.67 118.06 0.00 0.00 4220.39 1359.37 5752.29 00:17:19.173 00:17:19.173 real 0m1.742s 00:17:19.173 user 0m1.383s 00:17:19.173 sys 0m0.237s 00:17:19.173 16:55:40 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:19.173 16:55:40 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:17:19.173 ************************************ 00:17:19.173 END TEST bdev_write_zeroes 00:17:19.173 ************************************ 00:17:19.432 16:55:40 blockdev_raid5f -- bdev/blockdev.sh@781 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:19.433 16:55:40 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:17:19.433 16:55:40 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:19.433 16:55:40 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:19.433 ************************************ 00:17:19.433 START TEST bdev_json_nonenclosed 00:17:19.433 ************************************ 00:17:19.433 16:55:40 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:19.433 [2024-09-29 16:55:40.986547] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:17:19.433 [2024-09-29 16:55:40.986689] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100554 ] 00:17:19.692 [2024-09-29 16:55:41.134331] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:19.692 [2024-09-29 16:55:41.189344] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:19.692 [2024-09-29 16:55:41.189469] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:17:19.692 [2024-09-29 16:55:41.189492] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:19.692 [2024-09-29 16:55:41.189507] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:19.692 00:17:19.692 real 0m0.409s 00:17:19.692 user 0m0.176s 00:17:19.692 sys 0m0.129s 00:17:19.692 16:55:41 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:19.692 16:55:41 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:17:19.692 ************************************ 00:17:19.692 END TEST bdev_json_nonenclosed 00:17:19.692 ************************************ 00:17:19.951 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@784 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:19.951 16:55:41 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:17:19.951 16:55:41 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:19.951 16:55:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:19.951 ************************************ 00:17:19.951 START TEST bdev_json_nonarray 00:17:19.951 ************************************ 00:17:19.951 16:55:41 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:19.951 [2024-09-29 16:55:41.465187] Starting SPDK v25.01-pre git sha1 09cc66129 / DPDK 22.11.4 initialization... 00:17:19.951 [2024-09-29 16:55:41.465316] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100574 ] 00:17:19.951 [2024-09-29 16:55:41.611302] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:20.210 [2024-09-29 16:55:41.668391] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:20.210 [2024-09-29 16:55:41.668506] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:17:20.210 [2024-09-29 16:55:41.668529] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:17:20.210 [2024-09-29 16:55:41.668544] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:17:20.210 00:17:20.210 real 0m0.402s 00:17:20.210 user 0m0.170s 00:17:20.210 sys 0m0.128s 00:17:20.210 16:55:41 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:20.210 16:55:41 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:17:20.210 ************************************ 00:17:20.210 END TEST bdev_json_nonarray 00:17:20.210 ************************************ 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@786 -- # [[ raid5f == bdev ]] 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@793 -- # [[ raid5f == gpt ]] 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@797 -- # [[ raid5f == crypto_sw ]] 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@809 -- # trap - SIGINT SIGTERM EXIT 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@810 -- # cleanup 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:17:20.210 16:55:41 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:17:20.210 00:17:20.210 real 0m34.846s 00:17:20.210 user 0m47.450s 00:17:20.210 sys 0m4.670s 00:17:20.210 16:55:41 blockdev_raid5f -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:20.210 16:55:41 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:20.210 ************************************ 00:17:20.210 END TEST blockdev_raid5f 00:17:20.210 ************************************ 00:17:20.469 16:55:41 -- spdk/autotest.sh@194 -- # uname -s 00:17:20.469 16:55:41 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:17:20.469 16:55:41 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:20.469 16:55:41 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:17:20.469 16:55:41 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@252 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@256 -- # timing_exit lib 00:17:20.469 16:55:41 -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:20.469 16:55:41 -- common/autotest_common.sh@10 -- # set +x 00:17:20.469 16:55:41 -- spdk/autotest.sh@258 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@263 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@272 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@307 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@320 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@329 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@334 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@351 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:17:20.469 16:55:41 -- spdk/autotest.sh@362 -- # [[ 0 -eq 1 ]] 00:17:20.469 16:55:41 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:17:20.469 16:55:41 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:17:20.469 16:55:41 -- spdk/autotest.sh@374 -- # [[ '' -eq 1 ]] 00:17:20.469 16:55:41 -- spdk/autotest.sh@381 -- # trap - SIGINT SIGTERM EXIT 00:17:20.469 16:55:41 -- spdk/autotest.sh@383 -- # timing_enter post_cleanup 00:17:20.469 16:55:41 -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:20.469 16:55:41 -- common/autotest_common.sh@10 -- # set +x 00:17:20.469 16:55:41 -- spdk/autotest.sh@384 -- # autotest_cleanup 00:17:20.469 16:55:41 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:17:20.469 16:55:41 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:17:20.469 16:55:41 -- common/autotest_common.sh@10 -- # set +x 00:17:23.010 INFO: APP EXITING 00:17:23.010 INFO: killing all VMs 00:17:23.010 INFO: killing vhost app 00:17:23.010 INFO: EXIT DONE 00:17:23.271 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:23.271 Waiting for block devices as requested 00:17:23.271 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:17:23.531 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:17:24.472 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:17:24.472 Cleaning 00:17:24.472 Removing: /var/run/dpdk/spdk0/config 00:17:24.472 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:17:24.472 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:17:24.472 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:17:24.472 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:17:24.472 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:17:24.472 Removing: /var/run/dpdk/spdk0/hugepage_info 00:17:24.472 Removing: /dev/shm/spdk_tgt_trace.pid68812 00:17:24.472 Removing: /var/run/dpdk/spdk0 00:17:24.472 Removing: /var/run/dpdk/spdk_pid100189 00:17:24.472 Removing: /var/run/dpdk/spdk_pid100353 00:17:24.472 Removing: /var/run/dpdk/spdk_pid100430 00:17:24.472 Removing: /var/run/dpdk/spdk_pid100512 00:17:24.472 Removing: /var/run/dpdk/spdk_pid100554 00:17:24.472 Removing: /var/run/dpdk/spdk_pid100574 00:17:24.472 Removing: /var/run/dpdk/spdk_pid68643 00:17:24.472 Removing: /var/run/dpdk/spdk_pid68812 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69013 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69101 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69124 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69241 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69259 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69441 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69515 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69600 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69700 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69786 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69820 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69862 00:17:24.472 Removing: /var/run/dpdk/spdk_pid69928 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70050 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70469 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70521 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70569 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70585 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70654 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70664 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70735 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70746 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70799 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70817 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70859 00:17:24.472 Removing: /var/run/dpdk/spdk_pid70877 00:17:24.472 Removing: /var/run/dpdk/spdk_pid71004 00:17:24.472 Removing: /var/run/dpdk/spdk_pid71045 00:17:24.472 Removing: /var/run/dpdk/spdk_pid71124 00:17:24.472 Removing: /var/run/dpdk/spdk_pid72295 00:17:24.472 Removing: /var/run/dpdk/spdk_pid72490 00:17:24.472 Removing: /var/run/dpdk/spdk_pid72619 00:17:24.472 Removing: /var/run/dpdk/spdk_pid73229 00:17:24.733 Removing: /var/run/dpdk/spdk_pid73424 00:17:24.733 Removing: /var/run/dpdk/spdk_pid73553 00:17:24.733 Removing: /var/run/dpdk/spdk_pid74158 00:17:24.733 Removing: /var/run/dpdk/spdk_pid74477 00:17:24.733 Removing: /var/run/dpdk/spdk_pid74606 00:17:24.733 Removing: /var/run/dpdk/spdk_pid75947 00:17:24.733 Removing: /var/run/dpdk/spdk_pid76184 00:17:24.733 Removing: /var/run/dpdk/spdk_pid76319 00:17:24.733 Removing: /var/run/dpdk/spdk_pid77654 00:17:24.733 Removing: /var/run/dpdk/spdk_pid77891 00:17:24.733 Removing: /var/run/dpdk/spdk_pid78031 00:17:24.733 Removing: /var/run/dpdk/spdk_pid79361 00:17:24.733 Removing: /var/run/dpdk/spdk_pid79790 00:17:24.733 Removing: /var/run/dpdk/spdk_pid79919 00:17:24.733 Removing: /var/run/dpdk/spdk_pid81348 00:17:24.733 Removing: /var/run/dpdk/spdk_pid81592 00:17:24.733 Removing: /var/run/dpdk/spdk_pid81726 00:17:24.733 Removing: /var/run/dpdk/spdk_pid83151 00:17:24.733 Removing: /var/run/dpdk/spdk_pid83398 00:17:24.733 Removing: /var/run/dpdk/spdk_pid83528 00:17:24.733 Removing: /var/run/dpdk/spdk_pid84953 00:17:24.733 Removing: /var/run/dpdk/spdk_pid85423 00:17:24.733 Removing: /var/run/dpdk/spdk_pid85558 00:17:24.733 Removing: /var/run/dpdk/spdk_pid85685 00:17:24.733 Removing: /var/run/dpdk/spdk_pid86084 00:17:24.733 Removing: /var/run/dpdk/spdk_pid86792 00:17:24.733 Removing: /var/run/dpdk/spdk_pid87154 00:17:24.733 Removing: /var/run/dpdk/spdk_pid87831 00:17:24.733 Removing: /var/run/dpdk/spdk_pid88255 00:17:24.733 Removing: /var/run/dpdk/spdk_pid88991 00:17:24.733 Removing: /var/run/dpdk/spdk_pid89383 00:17:24.733 Removing: /var/run/dpdk/spdk_pid91308 00:17:24.733 Removing: /var/run/dpdk/spdk_pid91735 00:17:24.733 Removing: /var/run/dpdk/spdk_pid92159 00:17:24.733 Removing: /var/run/dpdk/spdk_pid94197 00:17:24.733 Removing: /var/run/dpdk/spdk_pid94670 00:17:24.733 Removing: /var/run/dpdk/spdk_pid95152 00:17:24.733 Removing: /var/run/dpdk/spdk_pid96191 00:17:24.733 Removing: /var/run/dpdk/spdk_pid96503 00:17:24.733 Removing: /var/run/dpdk/spdk_pid97425 00:17:24.733 Removing: /var/run/dpdk/spdk_pid97732 00:17:24.733 Removing: /var/run/dpdk/spdk_pid98649 00:17:24.733 Removing: /var/run/dpdk/spdk_pid98961 00:17:24.733 Removing: /var/run/dpdk/spdk_pid99631 00:17:24.733 Removing: /var/run/dpdk/spdk_pid99886 00:17:24.733 Removing: /var/run/dpdk/spdk_pid99926 00:17:24.733 Removing: /var/run/dpdk/spdk_pid99957 00:17:24.733 Clean 00:17:24.994 16:55:46 -- common/autotest_common.sh@1451 -- # return 0 00:17:24.994 16:55:46 -- spdk/autotest.sh@385 -- # timing_exit post_cleanup 00:17:24.994 16:55:46 -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:24.994 16:55:46 -- common/autotest_common.sh@10 -- # set +x 00:17:24.994 16:55:46 -- spdk/autotest.sh@387 -- # timing_exit autotest 00:17:24.994 16:55:46 -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:24.994 16:55:46 -- common/autotest_common.sh@10 -- # set +x 00:17:24.994 16:55:46 -- spdk/autotest.sh@388 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:17:24.994 16:55:46 -- spdk/autotest.sh@390 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:17:24.994 16:55:46 -- spdk/autotest.sh@390 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:17:24.994 16:55:46 -- spdk/autotest.sh@392 -- # [[ y == y ]] 00:17:24.994 16:55:46 -- spdk/autotest.sh@394 -- # hostname 00:17:24.994 16:55:46 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:17:25.254 geninfo: WARNING: invalid characters removed from testname! 00:17:47.207 16:56:07 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:49.114 16:56:10 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:51.023 16:56:12 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:53.565 16:56:14 -- spdk/autotest.sh@401 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:55.476 16:56:17 -- spdk/autotest.sh@402 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:58.050 16:56:19 -- spdk/autotest.sh@403 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:17:59.959 16:56:21 -- spdk/autotest.sh@404 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:18:00.220 16:56:21 -- common/autotest_common.sh@1680 -- $ [[ y == y ]] 00:18:00.220 16:56:21 -- common/autotest_common.sh@1681 -- $ lcov --version 00:18:00.220 16:56:21 -- common/autotest_common.sh@1681 -- $ awk '{print $NF}' 00:18:00.220 16:56:21 -- common/autotest_common.sh@1681 -- $ lt 1.15 2 00:18:00.220 16:56:21 -- scripts/common.sh@373 -- $ cmp_versions 1.15 '<' 2 00:18:00.220 16:56:21 -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:18:00.220 16:56:21 -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:18:00.220 16:56:21 -- scripts/common.sh@336 -- $ IFS=.-: 00:18:00.220 16:56:21 -- scripts/common.sh@336 -- $ read -ra ver1 00:18:00.220 16:56:21 -- scripts/common.sh@337 -- $ IFS=.-: 00:18:00.220 16:56:21 -- scripts/common.sh@337 -- $ read -ra ver2 00:18:00.220 16:56:21 -- scripts/common.sh@338 -- $ local 'op=<' 00:18:00.220 16:56:21 -- scripts/common.sh@340 -- $ ver1_l=2 00:18:00.220 16:56:21 -- scripts/common.sh@341 -- $ ver2_l=1 00:18:00.220 16:56:21 -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:18:00.220 16:56:21 -- scripts/common.sh@344 -- $ case "$op" in 00:18:00.220 16:56:21 -- scripts/common.sh@345 -- $ : 1 00:18:00.220 16:56:21 -- scripts/common.sh@364 -- $ (( v = 0 )) 00:18:00.220 16:56:21 -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:00.220 16:56:21 -- scripts/common.sh@365 -- $ decimal 1 00:18:00.220 16:56:21 -- scripts/common.sh@353 -- $ local d=1 00:18:00.220 16:56:21 -- scripts/common.sh@354 -- $ [[ 1 =~ ^[0-9]+$ ]] 00:18:00.220 16:56:21 -- scripts/common.sh@355 -- $ echo 1 00:18:00.220 16:56:21 -- scripts/common.sh@365 -- $ ver1[v]=1 00:18:00.220 16:56:21 -- scripts/common.sh@366 -- $ decimal 2 00:18:00.220 16:56:21 -- scripts/common.sh@353 -- $ local d=2 00:18:00.220 16:56:21 -- scripts/common.sh@354 -- $ [[ 2 =~ ^[0-9]+$ ]] 00:18:00.220 16:56:21 -- scripts/common.sh@355 -- $ echo 2 00:18:00.220 16:56:21 -- scripts/common.sh@366 -- $ ver2[v]=2 00:18:00.220 16:56:21 -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:18:00.220 16:56:21 -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:18:00.220 16:56:21 -- scripts/common.sh@368 -- $ return 0 00:18:00.220 16:56:21 -- common/autotest_common.sh@1682 -- $ lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:00.220 16:56:21 -- common/autotest_common.sh@1694 -- $ export 'LCOV_OPTS= 00:18:00.220 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.220 --rc genhtml_branch_coverage=1 00:18:00.220 --rc genhtml_function_coverage=1 00:18:00.220 --rc genhtml_legend=1 00:18:00.220 --rc geninfo_all_blocks=1 00:18:00.220 --rc geninfo_unexecuted_blocks=1 00:18:00.220 00:18:00.220 ' 00:18:00.220 16:56:21 -- common/autotest_common.sh@1694 -- $ LCOV_OPTS=' 00:18:00.220 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.220 --rc genhtml_branch_coverage=1 00:18:00.220 --rc genhtml_function_coverage=1 00:18:00.220 --rc genhtml_legend=1 00:18:00.220 --rc geninfo_all_blocks=1 00:18:00.220 --rc geninfo_unexecuted_blocks=1 00:18:00.220 00:18:00.220 ' 00:18:00.220 16:56:21 -- common/autotest_common.sh@1695 -- $ export 'LCOV=lcov 00:18:00.220 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.220 --rc genhtml_branch_coverage=1 00:18:00.220 --rc genhtml_function_coverage=1 00:18:00.220 --rc genhtml_legend=1 00:18:00.220 --rc geninfo_all_blocks=1 00:18:00.220 --rc geninfo_unexecuted_blocks=1 00:18:00.220 00:18:00.220 ' 00:18:00.220 16:56:21 -- common/autotest_common.sh@1695 -- $ LCOV='lcov 00:18:00.220 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:00.220 --rc genhtml_branch_coverage=1 00:18:00.221 --rc genhtml_function_coverage=1 00:18:00.221 --rc genhtml_legend=1 00:18:00.221 --rc geninfo_all_blocks=1 00:18:00.221 --rc geninfo_unexecuted_blocks=1 00:18:00.221 00:18:00.221 ' 00:18:00.221 16:56:21 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:18:00.221 16:56:21 -- scripts/common.sh@15 -- $ shopt -s extglob 00:18:00.221 16:56:21 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:18:00.221 16:56:21 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:00.221 16:56:21 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:00.221 16:56:21 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:00.221 16:56:21 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:00.221 16:56:21 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:00.221 16:56:21 -- paths/export.sh@5 -- $ export PATH 00:18:00.221 16:56:21 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:00.221 16:56:21 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:18:00.221 16:56:21 -- common/autobuild_common.sh@479 -- $ date +%s 00:18:00.221 16:56:21 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1727628981.XXXXXX 00:18:00.221 16:56:21 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1727628981.UMtUh4 00:18:00.221 16:56:21 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:18:00.221 16:56:21 -- common/autobuild_common.sh@485 -- $ '[' -n v22.11.4 ']' 00:18:00.221 16:56:21 -- common/autobuild_common.sh@486 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:18:00.221 16:56:21 -- common/autobuild_common.sh@486 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:18:00.221 16:56:21 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:18:00.221 16:56:21 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:18:00.221 16:56:21 -- common/autobuild_common.sh@495 -- $ get_config_params 00:18:00.221 16:56:21 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:18:00.221 16:56:21 -- common/autotest_common.sh@10 -- $ set +x 00:18:00.221 16:56:21 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:18:00.221 16:56:21 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:18:00.221 16:56:21 -- pm/common@17 -- $ local monitor 00:18:00.221 16:56:21 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:00.221 16:56:21 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:00.221 16:56:21 -- pm/common@25 -- $ sleep 1 00:18:00.221 16:56:21 -- pm/common@21 -- $ date +%s 00:18:00.221 16:56:21 -- pm/common@21 -- $ date +%s 00:18:00.221 16:56:21 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1727628981 00:18:00.221 16:56:21 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1727628981 00:18:00.221 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1727628981_collect-vmstat.pm.log 00:18:00.221 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1727628981_collect-cpu-load.pm.log 00:18:01.161 16:56:22 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:18:01.161 16:56:22 -- spdk/autopackage.sh@10 -- $ [[ 0 -eq 1 ]] 00:18:01.161 16:56:22 -- spdk/autopackage.sh@14 -- $ timing_finish 00:18:01.161 16:56:22 -- common/autotest_common.sh@736 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:18:01.161 16:56:22 -- common/autotest_common.sh@737 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:18:01.161 16:56:22 -- common/autotest_common.sh@740 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:01.422 16:56:22 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:18:01.422 16:56:22 -- pm/common@29 -- $ signal_monitor_resources TERM 00:18:01.422 16:56:22 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:18:01.422 16:56:22 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:01.422 16:56:22 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:18:01.422 16:56:22 -- pm/common@44 -- $ pid=102084 00:18:01.422 16:56:22 -- pm/common@50 -- $ kill -TERM 102084 00:18:01.422 16:56:22 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:01.422 16:56:22 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:18:01.422 16:56:22 -- pm/common@44 -- $ pid=102086 00:18:01.422 16:56:22 -- pm/common@50 -- $ kill -TERM 102086 00:18:01.422 + [[ -n 6155 ]] 00:18:01.422 + sudo kill 6155 00:18:01.432 [Pipeline] } 00:18:01.448 [Pipeline] // timeout 00:18:01.454 [Pipeline] } 00:18:01.468 [Pipeline] // stage 00:18:01.474 [Pipeline] } 00:18:01.489 [Pipeline] // catchError 00:18:01.498 [Pipeline] stage 00:18:01.501 [Pipeline] { (Stop VM) 00:18:01.513 [Pipeline] sh 00:18:01.797 + vagrant halt 00:18:04.337 ==> default: Halting domain... 00:18:12.480 [Pipeline] sh 00:18:12.763 + vagrant destroy -f 00:18:15.305 ==> default: Removing domain... 00:18:15.318 [Pipeline] sh 00:18:15.603 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:18:15.614 [Pipeline] } 00:18:15.629 [Pipeline] // stage 00:18:15.634 [Pipeline] } 00:18:15.649 [Pipeline] // dir 00:18:15.654 [Pipeline] } 00:18:15.668 [Pipeline] // wrap 00:18:15.674 [Pipeline] } 00:18:15.687 [Pipeline] // catchError 00:18:15.696 [Pipeline] stage 00:18:15.698 [Pipeline] { (Epilogue) 00:18:15.712 [Pipeline] sh 00:18:15.997 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:18:20.207 [Pipeline] catchError 00:18:20.209 [Pipeline] { 00:18:20.222 [Pipeline] sh 00:18:20.511 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:18:20.511 Artifacts sizes are good 00:18:20.520 [Pipeline] } 00:18:20.532 [Pipeline] // catchError 00:18:20.541 [Pipeline] archiveArtifacts 00:18:20.548 Archiving artifacts 00:18:20.682 [Pipeline] cleanWs 00:18:20.696 [WS-CLEANUP] Deleting project workspace... 00:18:20.696 [WS-CLEANUP] Deferred wipeout is used... 00:18:20.724 [WS-CLEANUP] done 00:18:20.726 [Pipeline] } 00:18:20.742 [Pipeline] // stage 00:18:20.747 [Pipeline] } 00:18:20.761 [Pipeline] // node 00:18:20.766 [Pipeline] End of Pipeline 00:18:20.802 Finished: SUCCESS